Abstract
Resilience is a major roadblock for HPC executions on future exascale systems. These systems will typically gather millions of CPU cores running up to a billion threads. Projections from current large systems and technology evolution predict errors will happen in exascale systems many times per day. These errors will propagate and generate various kinds of malfunctions, from simple process crashes to result corruptions. The past five years have seen extraordinary technical progress in many domains related to exascale resilience. Several technical options, initially considered inapplicable or unrealistic in the HPC context, have demonstrated surprising successes. Despite this progress, the exascale resilience problem is not solved, and the community is still facing the difficult challenge of ensuring that exascale applications complete and generate correct results while running on unstable systems. Since 2009, many workshops, studies, and reports have improved the definition of the resilience problem and provided refined recommendations. Some projections made during the previous decades and some priorities established from these projections need to be revised. This paper surveys what the community has learned in the past five years and summarizes the research problems still considered critical by the HPC community.
Original language | English |
---|---|
Pages (from-to) | 4-27 |
Number of pages | 24 |
Journal | Supercomputing Frontiers and Innovations |
Volume | 1 |
Issue number | 1 |
DOIs | |
State | Published - 2014 |
Funding
We thank Estaban Meneses for his help in the software section. This work was supported by the U.S. Department of Energy, Office of Science, Advanced Scientific Computing Research, under Contract DE-AC02-06CH11357, and under Award DESC0004131. This work was also supported in part by the Office of Advanced Scientific Computing Research, Office of Science, U.S. Department of Energy award DE-FG02-13ER26138/DE-SC0010049. This research is part of the Blue Waters sustained-petascale computing project, which is supported by the National Science Foundation (award number ACI 1238993) and the state of Illinois. Blue Waters is a joint effort of the University of Illinois at Urbana-Champaign and its National Center for Supercomputing Applications. The submitted manuscript has been created by UChicago Argonne, LLC, Operator of Argonne National Laboratory ('Argonne'). Argonne, a U.S. Department of Energy Office of Science laboratory, is operated under Contract No. DE-AC02-06CH11357. The U.S. Government retains for itself, and others acting on its behalf, a paid-up nonexclusive, irrevocable worldwide license in said article to reproduce, prepare derivative works, distribute copies to the public, and perform publicly and display publicly, by or on behalf of the Government.
Funders | Funder number |
---|---|
U.S. Department of Energy Office of Science Laboratory | |
National Science Foundation | ACI 1238993 |
U.S. Department of Energy | |
University of Illinois at Urbana-Champaign | |
Office of Science | |
Advanced Scientific Computing Research | DESC0004131, DE-AC02-06CH11357, DE-FG02-13ER26138/DE-SC0010049 |
Argonne National Laboratory | |
University of Chicago | |
University of Illinois | |
National Centre for Supercomputing Applications | |
National Science Foundation |
Keywords
- Exascale
- Fault-tolerance techniques
- Resilience