Optimizing deep learning hyper-parameters through an evolutionary algorithm

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

338 Scopus citations

Abstract

There has been a recent surge of success in utilizing Deep Learning (DL) in imaging and speech applications for its relatively automatic feature generation and, in particular for convolutional neural networks (CNNs), high accuracy classification abilities. While these models learn their parameters through data-driven methods, model selection (as architecture construction) through hyper-parameter choices remains a tedious and highly intuition driven task. To address this, Multi-node Evolutionary Neural Networks for Deep Learning (MENNDL) is proposed as a method for automating network selection on computational clusters through hyper-parameter optimization performed via genetic algorithms.

Original languageEnglish
Title of host publicationProceedings of MLHPC 2015
Subtitle of host publicationMachine Learning in High-Performance Computing Environments - Held in conjunction with SC 2015: The International Conference for High Performance Computing, Networking, Storage and Analysis
PublisherAssociation for Computing Machinery, Inc
ISBN (Electronic)9781450340069
DOIs
StatePublished - Nov 15 2015
EventWorkshop on Machine Learning in High-Performance Computing Environments, MLHPC 2015 - Austin, United States
Duration: Nov 15 2015 → …

Publication series

NameProceedings of MLHPC 2015: Machine Learning in High-Performance Computing Environments - Held in conjunction with SC 2015: The International Conference for High Performance Computing, Networking, Storage and Analysis

Conference

ConferenceWorkshop on Machine Learning in High-Performance Computing Environments, MLHPC 2015
Country/TerritoryUnited States
CityAustin
Period11/15/15 → …

Funding

This manuscript has been authored by UT-Battelle, LLC under Contract No. DE-AC05-00OR22725 with the U.S. Department of Energy. The United States Government retains and the publisher, by accepting the article for publication, acknowledges that the United States Government retains a non-exclusive, paid-up, irrevocable, world-wide license to publish or reproduce the published form of this manuscript, or allow others to do so, for United States Government purposes. The Department of Energy will provide public access to these results of federally sponsored research in accordance with the DOE Public Access Plan(http://energy.gov/downloads/doe-public-access-plan). This research used resources of the Oak Ridge Leadership Computing Facility at the Oak Ridge National Laboratory, which is supported by the Office of Science of the U.S. Department of Energy under Contract No. DE-AC05-00OR22725.

Keywords

  • Convolutional neural networks
  • Deep learning
  • Evolutionary algorithm
  • Hyper-parameter optimization

Fingerprint

Dive into the research topics of 'Optimizing deep learning hyper-parameters through an evolutionary algorithm'. Together they form a unique fingerprint.

Cite this