TY - GEN
T1 - Towards Rapid Autonomous Electron Microscopy with Active Meta-Learning
AU - Saranathan, Gayathri
AU - Foltin, Martin
AU - Tripathy, Aalap
AU - Ziatdinov, Maxim
AU - Koomthanam, Ann Mary Justine
AU - Bhattacharya, Suparna
AU - Ghosh, Ayana
AU - Roccapriore, Kevin
AU - Sukumar, Sreenivas Rangan
AU - Faraboschi, Paolo
N1 - Publisher Copyright:
© 2023 ACM.
PY - 2023/11/12
Y1 - 2023/11/12
N2 - We introduce a novel approach, Active Meta-learning, to improve computational control across various scientific experiments. It's particularly valuable for spectral reconstruction in STEM EELS nanoparticle plasmonic images. Traditionally, separate AI models were trained for each experiment via active learning, but this approach could face scalability issues with high-resolution data and the need for complex AI models due to intricate structure-property relationships. In this work we demonstrate the feasibility of learning AI structural representations across multiple experiments. We train a meta model from 10 prior experiments carried out such that the model can adapt to new unseen conditions in considerably less time than when trained from scratch. We utilize the Reptile algorithm, a first-order, model-agnostic meta-learning approach. To enhance and expand the meta-training dataset, conventional computer vision methods are applied to augment images from previous experiments. We observe up to ~30-40% reduction in the number of training epochs for active learning exploration. The approach will be extended to distributed meta-learning workflows; meta-model trained in HPC datacenter using data from different microscopy sites and pushed to individual sites for active learning.
AB - We introduce a novel approach, Active Meta-learning, to improve computational control across various scientific experiments. It's particularly valuable for spectral reconstruction in STEM EELS nanoparticle plasmonic images. Traditionally, separate AI models were trained for each experiment via active learning, but this approach could face scalability issues with high-resolution data and the need for complex AI models due to intricate structure-property relationships. In this work we demonstrate the feasibility of learning AI structural representations across multiple experiments. We train a meta model from 10 prior experiments carried out such that the model can adapt to new unseen conditions in considerably less time than when trained from scratch. We utilize the Reptile algorithm, a first-order, model-agnostic meta-learning approach. To enhance and expand the meta-training dataset, conventional computer vision methods are applied to augment images from previous experiments. We observe up to ~30-40% reduction in the number of training epochs for active learning exploration. The approach will be extended to distributed meta-learning workflows; meta-model trained in HPC datacenter using data from different microscopy sites and pushed to individual sites for active learning.
KW - Active Learning
KW - BO
KW - DKL
KW - DKLGPR
KW - EELS
KW - MAML
KW - Meta-learning
KW - Reptile
KW - STEM
UR - http://www.scopus.com/inward/record.url?scp=85178164753&partnerID=8YFLogxK
U2 - 10.1145/3624062.3626085
DO - 10.1145/3624062.3626085
M3 - Conference contribution
AN - SCOPUS:85178164753
T3 - ACM International Conference Proceeding Series
SP - 81
EP - 87
BT - Proceedings of 2023 SC Workshops of the International Conference on High Performance Computing, Network, Storage, and Analysis, SC Workshops 2023
PB - Association for Computing Machinery
T2 - 2023 International Conference on High Performance Computing, Network, Storage, and Analysis, SC Workshops 2023
Y2 - 12 November 2023 through 17 November 2023
ER -