A Vision for Coupling Operation of US Fusion Facilities with HPC Systems and the Implications for Workflows and Data Management

Sterling Smith, Emily Belli, Orso Meneghini, Reuben Budiardja, David Schissel, Jeff Candy, Tom Neiser, Adam Eubanks

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

2 Scopus citations

Abstract

The operation of large US Department of Energy (DOE) research facilities, like the DIII-D National Fusion Facility, results in the collection of complex multi-dimensional scientific datasets, both experimental and model-generated. In the future, it is envisioned that integrated data analysis coupled with large-scale high performance computing (HPC) simulations will be used to improve experimental planning and operation. Practically, massive data sets from these simulations provide the physics basis for generation of both reduced semi-analytic and machine-learning-based models. Storage of both HPC simulation datasets (generated from US DOE leadership computing facilities) and experimental datasets presents significant challenges. In this paper, we present a vision for a DOE-wide data management workflow that integrates US DOE fusion facilities with leadership computing facilities. Data persistence and long-term availability beyond the length of allocated projects is essential, particularly for verification and recalibration of artificial intelligence and machine learning (AI/ML) models. Because these data sets are often generated and shared among hundreds of users across multiple leadership computing facility centers, they would benefit from cross-platform accessibility, persistent identifiers (e.g. DOI, or digital object identifier), and provenance tracking. The ability to handle different data access patterns suggests that a combination of low cost, high latency (e.g. for storing ML training sets) and high cost, low latency systems (e.g. for real-time, integrated machine control feedback) may be needed.

Original languageEnglish
Title of host publicationAccelerating Science and Engineering Discoveries Through Integrated Research Infrastructure for Experiment, Big Data, Modeling and Simulation - 22nd Smoky Mountains Computational Sciences and Engineering Conference, SMC 2022, Revised Selected Papers
EditorsKothe Doug, Geist Al, Swaroop Pophale, Hong Liu, Suzanne Parete-Koon
PublisherSpringer Science and Business Media Deutschland GmbH
Pages87-100
Number of pages14
ISBN (Print)9783031236051
DOIs
StatePublished - 2022
EventSmoky Mountains Computational Sciences and Engineering Conference, SMC 2022 - Virtual, Online
Duration: Aug 24 2022Aug 25 2022

Publication series

NameCommunications in Computer and Information Science
Volume1690 CCIS
ISSN (Print)1865-0929
ISSN (Electronic)1865-0937

Conference

ConferenceSmoky Mountains Computational Sciences and Engineering Conference, SMC 2022
CityVirtual, Online
Period08/24/2208/25/22

Funding

Acknowledgments. This work was supported by the U.S. Department of Energy under awards DE-FC02-04ER54698, DE-FG02-95ER54309, DE-FC02-06ER54873, and DE-SC0017992. Some material is based upon work supported by the U.S. Department of Energy, Office of Science, Office of Fusion Energy Sciences, using the DIII-D National Fusion Facility, a DOE Office of Science user facility, under Award(s) DE-FC02-04ER54698. Computing resources were also provided by the National Energy Research Scientific Computing Center, which is an Office of Science User Facility supported under Contract DE-AC02-05CH11231. An award of computer time was also provided by the INCITE program. This research used resources of the Oak Ridge Leadership Computing Facility, which is an Office of Science User Facility supported under Contract DE-AC05-00OR22725. This report was prepared as an account of work sponsored by an agency of the United States Government. Neither the United States Government nor any agency thereof, nor any of their employees, makes any warranty, express or implied, or assumes any legal liability or responsibility for the accuracy, completeness, or usefulness of any information, apparatus, product, or process disclosed, or represents that its use would not infringe privately owned rights. Reference herein to any specific commercial product, process, or service by trade name, trademark, manufacturer, or otherwise, does not necessarily constitute or imply its endorsement, recommendation, or favoring by the United States Government or any agency thereof. The views and opinions of authors expressed herein do not necessarily state or reflect those of the United States Government or any agency thereof. In 2007, a database of over 300 nonlinear simulations [14] based on the GYRO gyrokinetic code was successfully used to calibrate the original, widely-used TGLF reduced transport model [4]. The subsequent speed and accuracy of TGLF enabled a revolution in the accuracy of modeling calculations. The original GYRO simulations incorporated only ion-scale physics, typically assumed a pure plasma (deuterium ions + electrons), and were representative of weakly-shaped core plasma parameters. Several new gyrokinetic database efforts are underway, including the US CGYRODB database at NERSC based on the CGYRO code [12], the US MGKDB [15] at NERSC based on the GENE code [16], and the European Gyro-Kinetic Database Project (GKDB) hosted at gkdb.org based on the GKW [17] and GENE [16] codes. One goal of the gyrokinetic database projects is to store not only the physics input paramaters and selected output data, but also all code-specific model and resolution parameters for reproducibility. The two US databases, CGYRODB and MGKDB, were established and are currently directly organized through projects funded by the US DOE SciDAC (Scientific Discovery through Advanced Computing) program, whose aim is to bring together physicists, computer scientists, and mathematicians to develop new computational methods for solving challenging scientific problems. These projects are AToM: Advanced Tokamak Modeling Environment (for CGYRODB) and Partnership for Multiscale Gyrokinetic (MGK) Turbulence (for MGKDB). Both CGYRODB and MGKDB are presently hosted by Community File System (CFS) storage space at NERSC, where most of the simulations are also run with computational time in support of the associated SciDAC grants. Since the present database consists mostly of smaller-scale nonlinear simulations (e.g. ion-scale or electron-scale spatial resolution) with only highly-distilled output data, the CFS project space is presently sufficient. For example, the directory storage quota for the CGYRODB at NERSC is only 4 TB. A typical dataset for a single CGYRO ion-scale dataset is on the order of 0.1 GB.

Keywords

  • Big data
  • Data management
  • Fusion
  • High performance computing

Fingerprint

Dive into the research topics of 'A Vision for Coupling Operation of US Fusion Facilities with HPC Systems and the Implications for Workflows and Data Management'. Together they form a unique fingerprint.

Cite this