Back to Search Start Over

Using open-science workflow tools to produce SCEC CyberShake physics-based probabilistic seismic hazard models

Authors :
Scott Callaghan
Philip J. Maechling
Fabio Silva
Mei-Hui Su
Kevin R. Milner
Robert W. Graves
Kim B. Olsen
Yifeng Cui
Karan Vahi
Albert Kottke
Christine A. Goulet
Ewa Deelman
Thomas H. Jordan
Yehuda Ben-Zion
Source :
Frontiers in High Performance Computing, Vol 2 (2024)
Publication Year :
2024
Publisher :
Frontiers Media S.A., 2024.

Abstract

The Statewide (formerly Southern) California Earthquake Center (SCEC) conducts multidisciplinary earthquake system science research that aims to develop predictive models of earthquake processes, and to produce accurate seismic hazard information that can improve societal preparedness and resiliency to earthquake hazards. As part of this program, SCEC has developed the CyberShake platform, which calculates physics-based probabilistic seismic hazard analysis (PSHA) models for regions with high-quality seismic velocity and fault models. The CyberShake platform implements a sophisticated computational workflow that includes over 15 individual codes written by 6 developers. These codes are heterogeneous, ranging from short-running high-throughput serial CPU codes to large, long-running, parallel GPU codes. Additionally, CyberShake simulation campaigns are computationally extensive, typically producing tens of terabytes of meaningful scientific data and metadata over several months of around-the-clock execution on leadership-class supercomputers. To meet the needs of the CyberShake platform, we have developed an extreme-scale workflow stack, including the Pegasus Workflow Management System, HTCondor, Globus, and custom tools. We present this workflow software stack and identify how the CyberShake platform and supporting tools enable us to meet a variety of challenges that come with large-scale simulations, such as automated remote job submission, data management, and verification and validation. This platform enabled us to perform our most recent simulation campaign, CyberShake Study 22.12, from December 2022 to April 2023. During this time, our workflow tools executed approximately 32,000 jobs, and used up to 73% of the Summit system at Oak Ridge Leadership Computing Facility. Our workflow tools managed about 2.5 PB of total temporary and output data, and automatically staged 19 million output files totaling 74 TB back to archival storage on the University of Southern California's Center for Advanced Research Computing systems, including file-based relational data and large binary files to efficiently store millions of simulated seismograms. CyberShake extreme-scale workflows have generated simulation-based probabilistic seismic hazard models that are being used by seismological, engineering, and governmental communities.

Details

Language :
English
ISSN :
28137337
Volume :
2
Database :
Directory of Open Access Journals
Journal :
Frontiers in High Performance Computing
Publication Type :
Academic Journal
Accession number :
edsdoj.30b76b50e3854374ae111c6c0f6ddbf0
Document Type :
article
Full Text :
https://doi.org/10.3389/fhpcp.2024.1360720