1. End-to-end I/O portfolio for the summit supercomputing ecosystem
- Author
-
George S. Markomanolis, Ross Miller, Dustin Leverman, Sarp Oral, Feiyi Wang, Christopher Brumgard, Scott Atchley, Christopher Zimmer, Sudharshan S. Vazhkudai, Verónica G. Vergara Larrea, and Jesse Hanley
- Subjects
Input/output ,File system ,geography ,TOP500 ,Summit ,geography.geographical_feature_category ,Computer science ,IOPS ,Transparency (human–computer interaction) ,computer.software_genre ,Supercomputer ,Mount ,Operating system ,IBM ,Layer (object-oriented design) ,computer - Abstract
The I/O subsystem for the Summit supercomputer, No. 1 on the Top500 list, and its ecosystem of analysis platforms is composed of two distinct layers, namely the in-system layer and the center-wide parallel file system layer (PFS), Spider 3. The in-system layer uses node-local SSDs and provides 26.7 TB/s for reads, 9.7 TB/s for writes, and 4.6 billion IOPS to Summit. The Spider 3 PFS layer uses IBM's Spectrum Scale™ and provides 2.5 TB/s and 2.6 million IOPS to Summit and other systems. While deploying them as two distinct layers was operationally efficient, it also presented usability challenges in terms of multiple mount points and lack of transparency in data movement. To address these challenges, we have developed novel end-to-end I/O solutions for the concerted use of the two storage layers. We present the I/O subsystem architecture, the end-to-end I/O solution space, their design considerations and our deployment experience.
- Published
- 2019
- Full Text
- View/download PDF