HyperSpace: Distributed Bayesian Hyperparameter Optimization

M. Todd Young, Jacob Hinkle, Arvind Ramanathan, Ramakrishnan Kannan

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

14 Scopus citations

Abstract

As machine learning models continue to increase in complexity, so does the potential number of free model parameters commonly known as hyperparameters. While there has been considerable progress toward finding optimal configurations of these hyperparameters, many optimization procedures are treated as black boxes. We believe optimization methods should not only return a set of optimized hyperparameters, but also give insight into the effects of model hyperparameter settings. To this end, we present HyperSpace, a parallel implementation of Bayesian sequential model-based optimization. HyperSpace leverages high performance computing (HPC) resources to better understand unknown, potentially non-convex hyperparameter search spaces. We show that it is possible to learn the dependencies between model hyperparameters through the optimization process. By partitioning large search spaces and running many optimization procedures in parallel, we also show that it is possible to discover families of good hyperparameter settings over a variety of models including unsupervised clustering, regression, and classification tasks.

Original languageEnglish
Title of host publicationProceedings - 2018 30th International Symposium on Computer Architecture and High Performance Computing, SBAC-PAD 2018
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages339-347
Number of pages9
ISBN (Electronic)9781538677698
DOIs
StatePublished - Jul 2 2018
Event30th International Symposium on Computer Architecture and High Performance Computing, SBAC-PAD 2018 - Lyon, France
Duration: Sep 24 2018Sep 27 2018

Publication series

NameProceedings - 2018 30th International Symposium on Computer Architecture and High Performance Computing, SBAC-PAD 2018

Conference

Conference30th International Symposium on Computer Architecture and High Performance Computing, SBAC-PAD 2018
Country/TerritoryFrance
CityLyon
Period09/24/1809/27/18

Funding

This work has been supported in part by the Joint Design of Advanced Computing Solutions for Cancer (JDACS4C) program established by the U.S. Department of Energy (DOE) and the National Cancer Institute (NCI) of the National Institutes of Health. This work was performed under the auspices of the U.S. Department of Energy by Argonne National Laboratory under Contract DE-AC02-06-CH11357,Lawrence Livermore National Laboratory under Contract DE-AC52-07NA27344,Los Alamos National Laboratory under Contract DE-AC5206NA25396, and Oak Ridge National Laboratory under Contract DE-AC05-00OR22725.This research was supported by the Exascale Computing Project (17-SC-20-SC), a collaborative effort of the U.S. Department of Energy Office of Science and the National Nuclear Security Administration. This research used resources of the Oak Ridge Leadership Computing Facility at the Oak Ridge National Laboratory.

FundersFunder number
U.S. Department of Energy Office of Science
National Institutes of Health
U.S. Department of Energy
National Cancer Institute
National Nuclear Security Administration
Argonne National LaboratoryDE-AC02-06-CH11357
Lawrence Livermore National LaboratoryDE-AC52-07NA27344
Oak Ridge National Laboratory17-SC-20-SC, DE-AC05-00OR22725
Los Alamos National LaboratoryDE-AC5206NA25396

    Keywords

    • Bayesian optimization
    • HPC
    • SMBO
    • parallel computing

    Fingerprint

    Dive into the research topics of 'HyperSpace: Distributed Bayesian Hyperparameter Optimization'. Together they form a unique fingerprint.

    Cite this