Abstract
Cyclic DARTS (CDARTS) is a Differentiable Architecture Search (DARTS)-based approach to neural architecture search (NAS) that uses a cyclic feedback mechanism to train search and evaluation networks concurrently. This training protocol aims to optimize the search process and evaluate the deep evaluation network comprised of discretized candidate operations. However, this approach introduces a loss function for the evaluation network dependent on the search network. The dissimilarity between the evaluation network's loss function used during the search and retraining phases results in a search network that is a sub-optimal proxy for the final evaluation network accessed during retraining. We present a revised approach that removes the dependency of the evaluation network weights upon those of the search network. In addition, we introduce a modified process for relaxing the search network's zero operations that allows these operations to be retained in the final evaluation networks.
Original language | English |
---|---|
Title of host publication | Proceedings - 21st IEEE International Conference on Machine Learning and Applications, ICMLA 2022 |
Editors | M. Arif Wani, Mehmed Kantardzic, Vasile Palade, Daniel Neagu, Longzhi Yang, Kit-Yan Chan |
Publisher | Institute of Electrical and Electronics Engineers Inc. |
Pages | 1055-1062 |
Number of pages | 8 |
ISBN (Electronic) | 9781665462839 |
DOIs | |
State | Published - 2022 |
Event | 21st IEEE International Conference on Machine Learning and Applications, ICMLA 2022 - Nassau, Bahamas Duration: Dec 12 2022 → Dec 14 2022 |
Publication series
Name | Proceedings - 21st IEEE International Conference on Machine Learning and Applications, ICMLA 2022 |
---|
Conference
Conference | 21st IEEE International Conference on Machine Learning and Applications, ICMLA 2022 |
---|---|
Country/Territory | Bahamas |
City | Nassau |
Period | 12/12/22 → 12/14/22 |
Funding
This research used resources of the Compute and Data Environment for Science (CADES) at the Oak Ridge National Laboratory, which is supported by the Office of Science of the U.S. Department of Energy under Contract No. DE-AC05-00OR22725. Notice: This manuscript has been authored by UT-Battelle, LLC, under contract DE-AC05-00OR22725 with the US Department of Energy (DOE). The US government retains and the publisher, by accepting the article for publication, acknowledges that the US government retains a nonexclusive, paid-up, irrevocable, worldwide license to publish or reproduce the published form of this manuscript, or allow others to do so, for US government purposes. DOE will provide public access to these results of federally sponsored research in accordance with the DOE Public Access Plan (http://energy.gov/downloads/doe-public-access-plan).
Keywords
- AutoML
- Deep Learning
- Neural Architecture Search
- Neural Networks