TY - GEN
T1 - Ramifications of Evolving Misbehaving Convolutional Neural Network Kernel and Batch Sizes
AU - Coletti, Mark
AU - Lunga, Dalton
AU - Berres, Anne
AU - Sanyal, Jibonananda
AU - Rose, Amy
N1 - Publisher Copyright:
© 2018 IEEE.
PY - 2018/7/2
Y1 - 2018/7/2
N2 - Deep-learners have many hyper-parameters including learning rate, batch size, kernel size - all playing a significant role toward estimating high quality models. Discovering useful hyper-parameter guidelines is an active area of research, though the state of the art generally uses a brute force, uniform grid approach or random search for finding ideal settings. We share the preliminary results of using an alternative approach to deep learner hyper-parameter tuning that uses an evolutionary algorithm to improve the accuracy of a deep-learner models used in satellite imagery building footprint detection. We found that the kernel and batch size hyper-parameters surprisingly differed from sizes arrived at via a brute force uniform grid approach. These differences suggest a novel role for evolutionary algorithms in determining the number of convolution layers, as well as smaller batch sizes in improving deep-learner models.
AB - Deep-learners have many hyper-parameters including learning rate, batch size, kernel size - all playing a significant role toward estimating high quality models. Discovering useful hyper-parameter guidelines is an active area of research, though the state of the art generally uses a brute force, uniform grid approach or random search for finding ideal settings. We share the preliminary results of using an alternative approach to deep learner hyper-parameter tuning that uses an evolutionary algorithm to improve the accuracy of a deep-learner models used in satellite imagery building footprint detection. We found that the kernel and batch size hyper-parameters surprisingly differed from sizes arrived at via a brute force uniform grid approach. These differences suggest a novel role for evolutionary algorithms in determining the number of convolution layers, as well as smaller batch sizes in improving deep-learner models.
KW - Convolutional neural networks
KW - Deep learning
KW - Evolutionary algorithms
KW - Hyper-parameters
KW - Optimization
KW - Satellite imagery
KW - Settlement detection
UR - http://www.scopus.com/inward/record.url?scp=85063027712&partnerID=8YFLogxK
U2 - 10.1109/MLHPC.2018.8638644
DO - 10.1109/MLHPC.2018.8638644
M3 - Conference contribution
AN - SCOPUS:85063027712
T3 - Proceedings of MLHPC 2018: Machine Learning in HPC Environments, Held in conjunction with SC 2018: The International Conference for High Performance Computing, Networking, Storage and Analysis
SP - 106
EP - 113
BT - Proceedings of MLHPC 2018
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2018 IEEE/ACM Machine Learning in HPC Environments, MLHPC 2018
Y2 - 12 November 2018
ER -