Co-design Center for Exascale Machine Learning Technologies (ExaLearn)

Francis J. Alexander, James Ang, Jenna A. Bilbrey, Jan Balewski, Tiernan Casey, Ryan Chard, Jong Choi, Sutanay Choudhury, Bert Debusschere, Anthony M. DeGennaro, Nikoli Dryden, J. Austin Ellis, Ian Foster, Cristina Garcia Cardona, Sayan Ghosh, Peter Harrington, Yunzhi Huang, Shantenu Jha, Travis Johnston, Ai KagawaRamakrishnan Kannan, Neeraj Kumar, Zhengchun Liu, Naoya Maruyama, Satoshi Matsuoka, Erin McCarthy, Jamaludin Mohd-Yusof, Peter Nugent, Yosuke Oyama, Thomas Proffen, David Pugmire, Sivasankaran Rajamanickam, Vinay Ramakrishniah, Malachi Schram, Sudip K. Seal, Ganesh Sivaraman, Christine Sweeney, Li Tan, Rajeev Thakur, Brian Van Essen, Logan Ward, Paul Welch, Michael Wolf, Sotiris S. Xantheas, Kevin G. Yager, Shinjae Yoo, Byung Jun Yoon

Research output: Contribution to journalArticlepeer-review

15 Scopus citations

Abstract

Rapid growth in data, computational methods, and computing power is driving a remarkable revolution in what variously is termed machine learning (ML), statistical learning, computational learning, and artificial intelligence. In addition to highly visible successes in machine-based natural language translation, playing the game Go, and self-driving cars, these new technologies also have profound implications for computational and experimental science and engineering, as well as for the exascale computing systems that the Department of Energy (DOE) is developing to support those disciplines. Not only do these learning technologies open up exciting opportunities for scientific discovery on exascale systems, they also appear poised to have important implications for the design and use of exascale computers themselves, including high-performance computing (HPC) for ML and ML for HPC. The overarching goal of the ExaLearn co-design project is to provide exascale ML software for use by Exascale Computing Project (ECP) applications, other ECP co-design centers, and DOE experimental facilities and leadership class computing facilities.

Original languageEnglish
Pages (from-to)598-616
Number of pages19
JournalInternational Journal of High Performance Computing Applications
Volume35
Issue number6
DOIs
StatePublished - Nov 2021

Funding

The author(s) disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: This research is supported by the Exascale Computing Project (17-SC-20-SC), a collaborative effort of U.S. Department of Energy Office of Science and the National Nuclear Security Administration. A portion of the research was supported by JSPS KAKENHI Grant Number JP18J22858, Japan. The research has used resources of the Argonne and Oak Ridge Leadership Computing Facilities, Livermore Computing Facility, and Energy Research Scientific Computing Center (NERSC) DOE Office of Science User Facilities supported under Contracts DE-AC02-06CH11357, DE-AC05-00OR22725 and DE-AC52-07NA27344 (LLNL-JRNL-XXXXXX), DE-AC02-05CH11231, respectively.

FundersFunder number
LLNL-JRNL-XXXXXXDE-AC02-05CH11231
U.S. Department of Energy
Office of ScienceDE-AC05-00OR22725, DE-AC02-06CH11357, DE-AC52-07NA27344
Office of Science
National Nuclear Security Administration
Japan Society for the Promotion of ScienceJP18J22858
Japan Society for the Promotion of Science

    Keywords

    • Machine learning
    • active learning
    • exascale computing
    • high-performance computing for machine learning
    • machine learning for high-performance computing
    • reinforcement learning

    Fingerprint

    Dive into the research topics of 'Co-design Center for Exascale Machine Learning Technologies (ExaLearn)'. Together they form a unique fingerprint.

    Cite this