Burst buffer

From Wikipedia, the free encyclopedia
Jump to navigation Jump to search

In the high-performance computing environment, burst buffer is a fast and intermediate storage layer positioned between the front-end computing processes and the back-end storage systems. It emerges as a timely storage solution to bridge the ever-increasing performance gap between the processing speed of the compute nodes and the Input/output (I/O) bandwidth of the storage systems. Burst buffer is built from arrays of high-performance storage devices, such as NVRAM and SSD. It typically offers from one to two orders of magnitude higher I/O bandwidth than the back-end storage systems.

Use cases[edit]

The emergence of burst buffer fosters a wide variety of burst buffer based solutions that accelerate the scientific data movement on supercomputers. For example, scientific applications' life cycles typically alternate between computation phases and I/O phases.[1] Namely, after each round of computation (i.e., computation phase), all the computing processes concurrently write their intermediate data to the back-end storage systems (i.e., I/O phase), followed by another round of computation and data movement operations. With the deployment of burst buffer, processes can quickly write their data to burst buffer after one round of computation instead of writing to the slow hard disk based storage systems, and immediately proceed to the next round of computation without waiting for the data to be moved to the back-end storage systems;[2][3] the data are then asynchronously flushed from burst buffer to the storage systems at the same time with the next round of computation. In this way, the long I/O time spent in moving data to the storage systems is hidden behind the computation time. In addition, buffering data in burst buffer also gives applications plenty of opportunities to reshape the data traffic to the back-end storage systems for efficient bandwidth utilization of the storage systems.[4][5] In another common use case, scientific applications can stage their intermediate data in and out of burst buffer without interacting with the slower storage systems. Bypassing the storage systems allows applications to realize most of the performance benefit from burst buffer.[6]

Representative burst buffer architectures[edit]

There are two representative burst buffer architectures in the high-performance computing environment: node-local burst buffer and remote shared burst buffer. In the node-local burst buffer architecture, burst buffer storage is located on the individual compute node, so the aggregate burst buffer bandwidth grows linearly with the compute node count. This scalability benefit has been well-documented in recent literature.[7][8][9][10]It also comes with the demand for a scalable metadata management strategy to maintain a global namespace for data distributed across all the burst buffers.[11][12] In the remote shared burst buffer architecture, burst buffer storage resides on a fewer number of I/O nodes positioned between the compute nodes and the back-end storage systems. Data movement between the compute nodes and burst buffer needs to go through the network. Placing burst buffer on the I/O nodes facilitates the independent development, deployment and maintenance of the burst buffer service. Hence, several well-known commercialized software products have been developed to manage this type of burst buffer, such as DataWarp and Infinite Memory Engine. As the upcoming supercomputers are going to be deployed with multiple heterogeneous burst buffer layers, such as NVRAM on the compute nodes, and SSDs on the dedicated I/O nodes, there is a growing interest in designing and implementing a unified software solution that transparently moves data across multiple storage layers.[13][14][15]

Supercomputers deployed with burst buffer[edit]

Due to its importance, burst buffer has been widely deployed on the leadership-scale supercomputers. For example, node-local burst buffer has been installed on DASH supercomputer at the San Diego Supercomputer Center,[16] Tsubame supercomputers at Tokyo Institute of Technology, Theta and Aurora supercomputers at the Argonne National Laboratory, Summit supercomputer at the Oak Ridge National Laboratory, and Sierra supercomputer at the Lawrence Livermore National Laboratory, etc. Remote shared burst buffer has been adopted by Tianhe-2 supercomputer at the National Supercomputer Center in Guangzhou, Trinity supercomputer at the Los Alamos National Laboratory, and Cori supercomputer at the Lawrence Berkeley National Laboratory, etc.

References[edit]

  1. ^ "A Case of System-Wide Power Management for Scientific Applications" (PDF). IEEE. September 2013.
  2. ^ "BurstMem: A High-Performance Burst Buffer System for Scientific Applications" (PDF). IEEE. October 2014.
  3. ^ "On the Role of Burst Buffers in Leadership-Class Storage systems" (PDF). IEEE. April 2012.
  4. ^ "TRIO: Burst Buffer Based I/O Orchestration" (PDF). IEEE. September 2015.
  5. ^ "Leveraging Burst Buffer Coordination to Prevent I/O Interference" (PDF). IEEE. March 2017.
  6. ^ "An Ephemeral Burst-Buffer File System for Scientific Applications" (PDF). IEEE. November 2016.
  7. ^ "BurstFS: A Distributed Burst Buffer File System for Scientific Applications" (PDF). November 2015.
  8. ^ "Design, Modeling, and Evaluation of a Scalable Multi-level Checkpointing System" (pdf). ACM. November 2010.
  9. ^ "A 1 PB/s File System to Checkpoint Three Million MPI Tasks" (PDF). ACM. June 2013.
  10. ^ "FusionFS: Toward supporting data-intensive scientific applications on extreme-scale high-performance computing systems" (PDF). IEEE. October 2014.
  11. ^ "MetaKV: A Key-Value Store for Metadata Management of Distributed Burst Buffers" (PDF). IEEE. May 2017.
  12. ^ "ZHT: A Light-Weight Reliable Persistent Dynamic Scalable Zero-Hop Distributed Hash Table" (PDF). IEEE. May 2013.
  13. ^ "UniviStor: Integrated Hierarchical and Distributed Storage for HPC" (PDF). IEEE. Sep 2018.
  14. ^ "Hermes: a heterogeneous-aware multi-tiered distributed I/O buffering system" (PDF). ACM. June 2018.
  15. ^ "Toward Scalable and Asynchronous Object-centric Data Management for HPC" (PDF). IEEE. May 2018.
  16. ^ "DASH: a Recipe for a Flash-based Data Intensive Supercomputer" (PDF). ACM. November 2010.

External links[edit]