Abstract
Current deep learning approaches have been very successful using convolutional neural networks trained on large graphical-processing-unit-based computers. Three limitations of this approach are that (1) they are based on a simple layered network topology, i.e., highly connected layers, without intra-layer connections; (2) the networks are manually configured to achieve optimal results, and (3) the implementation of the network model is expensive in both cost and power. In this article, we evaluate deep learning models using three different computing architectures to address these problems: quantum computing to train complex topologies, high performance computing to automatically determine network topology, and neuromorphic computing for a low-power hardware implementation. We use the MNIST dataset for our experiment, due to input size limitations of current quantum computers. Our results show the feasibility of using the three architectures in tandem to address the above deep learning limitations.We show that a quantum computer can find high quality values of intra-layer connection weights in a tractable time as the complexity of the network increases, a high performance computer can find optimal layer-based topologies, and a neuromorphic computer can represent the complex topology and weights derived from the other architectures in low power memristive hardware.
Original language | English |
---|---|
Article number | 3178454 |
Journal | ACM Journal on Emerging Technologies in Computing Systems |
Volume | 14 |
Issue number | 2 |
DOIs | |
State | Published - Jul 2018 |
Funding
Notice: This manuscript was authored by UT-Battelle, LLC under contract DE-AC05-00OR22725 with the U.S. Department of Energy. The United States Government retains and the publisher, by accepting the article for publication, acknowledges that the United States Government retains a non-exclusive, paid-up, irrevocable, world-wide license to publish or reproduce the published form of this manuscript, or allow others to do so, for United States Government purposes. The Department of Energy will provide public access to these results of federally sponsored research in accordance with the DOE Public Access Plan (http://energy.gov/downloads/doe-public-access-plan). This material was based on work supported by the U.S. Department of Energy, Office of Science, Office of Advanced Scientific Computing Research, Robinson Pino, program manager, under contract DE-AC05-00OR22725. This research used resources of the Oak Ridge Leadership Computing Facility, which is a DOE Office of Science User Facility supported under contract DE-AC05-00OR22725. Authors’ addresses: T. E. Potok, C. Schuman, S. Young, and R. Patton, Computational Data Analytics Group, Oak Ridge National Laboratory, P.O. Box 2008, Oak Ridge, TN 37831; emails: {potokte, schumancd, youngsr, pattonrm}@ornl.gov; F. Spedalieri, J. Liu, and K.-T. Yao, University of Southern California, Information Sciences Institute, 4676 Admiralty Way, Suite 1001, Marina del Rey, CA 90292; emails: [email protected], [email protected], [email protected]; G. Rose and G. Chakma, Department of Electrical Engineering and Computer Science, University of Tennessee, 1520 Middle Dr, Suite 401, Knoxville, TN 37996 USA. Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]. © 2018 ACM 1550-4832/2018/07-ART19 $15.00 https://doi.org/10.1145/3178454
Funders | Funder number |
---|---|
DOE Office of Science | |
U.S. Department of Energy | |
Office of Science | |
Advanced Scientific Computing Research |