Scaling Laws of Graph Neural Networks for Atomistic Materials Modeling

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Atomistic materials modeling is a critical task with wide-ranging applications, from drug discovery to materials science, where accurate predictions of the target material property can lead to significant advancements in scientific discovery. Graph Neural Networks (GNNs) represent the state-of-the-art approach for modeling atomistic material data thanks to their capacity to capture complex relational structures. While machine learning performance has historically improved with larger models and datasets, GNNs for atomistic materials modeling remain relatively small compared to large language models (LLMs), which leverage billions of parameters and terabyte-scale datasets to achieve remarkable performance in their respective domains. To address this gap, we explore the scaling limits of GNNs for atomistic materials modeling by developing a foundational model with billions of parameters, trained on extensive datasets in terabytescale. Our approach incorporates techniques from LLM libraries to efficiently manage large-scale data and models, enabling both effective training and deployment of these large-scale GNN models. This work addresses three fundamental questions in scaling GNNs: the potential for scaling GNN model architectures, the effect of dataset size on model accuracy, and the applicability of LLM-inspired techniques to GNN architectures. Specifically, the outcomes of this study include (1) insights into the scaling laws for GNNs, highlighting the relationship between model size, dataset volume, and accuracy, (2) a foundational GNN model optimized for atomistic materials modeling, and (3) a GNN codebase enhanced with advanced LLM-based training techniques. Our findings lay the groundwork for large-scale GNNs with billions of parameters and terabyte-scale datasets, establishing a scalable pathway for future advancements in atomistic materials modeling.

Original languageEnglish
Title of host publication2025 62nd ACM/IEEE Design Automation Conference, DAC 2025
PublisherInstitute of Electrical and Electronics Engineers Inc.
ISBN (Electronic)9798331503048
DOIs
StatePublished - 2025
Event62nd ACM/IEEE Design Automation Conference, DAC 2025 - San Francisco, United States
Duration: Jun 22 2025Jun 25 2025

Publication series

NameProceedings - Design Automation Conference
ISSN (Print)0738-100X

Conference

Conference62nd ACM/IEEE Design Automation Conference, DAC 2025
Country/TerritoryUnited States
CitySan Francisco
Period06/22/2506/25/25

Funding

Massimiliano Lupo Pasini would like to thank Dr. Vladimir Protopopescu for his valuable feedback in the preparation of the manuscript. This research is sponsored by the Artificial Intelligence Initiative as part of the Laboratory Directed Research and Development (LDRD) Program of Oak Ridge National Laboratory, managed by UT-Battelle, LLC, for the US Department of Energy under contract DE-AC05-00OR22725. This work used resources of the Oak Ridge Leadership Computing Facility, which is supported by the Office of Science of the U.S. Department of Energy, under INCITE award CPH161. This work also used resources of the National Energy Research Scientific Computing Center, which is supported by the Office of Science of the U.S. Department of Energy under Contract No. DE-AC02-05CH11231, under awards ERCAP0027259 and ERCAP0030519. This work was supported in part by the National Science Foundation (NSF) (Award ID: 2400511 and 2016727), the Department of Health and Human Services Advanced Research Projects Agency for Health (ARPA-H) under Award Number AY1AX000003, and CoCoSys, one of the seven centers in JUMP 2.0, a Semiconductor Research Corporation (SRC) program sponsored by DARPA.

Keywords

  • AI/ML Application and Infrastructure
  • AI/ML System and Platform Design

Fingerprint

Dive into the research topics of 'Scaling Laws of Graph Neural Networks for Atomistic Materials Modeling'. Together they form a unique fingerprint.

Cite this