GPU Direct I/O with HDF5
Exascale HPC systems are being designed with accelerators, such as GPUs, to accelerate parts of applications. In machine learning workloads as well as large-scale simulations that use GPUs as accelerators, the CPU (or host) memory is currently used as a buffer for data transfers between GPU (or device) memory and the file system. If the CPU does not need to operate on the data, then this is sub-optimal because it wastes host memory by reserving space for duplicated data. Furthermore, this “bounce buffer” approach wastes CPU cycles spent on transferring data. A new technique, NVIDIA GPUDirect Storage (GDS), can eliminate the need to use the host memory as a bounce buffer. Thereby, it becomes possible to transfer data directly between the device memory and the file system. This direct data path shortens latency by omitting the extra copy and enables higher-bandwidth. To take full advantage of GDS in existing applications, it is necessary to provide support with existing I/O libraries, such as HDF5 and MPI-IO, which are heavily used in applications. In this paper, we describe our effort of integrating GDS with HDF5, the top I/O library at NERSC and at DOE leadership computing facilities. We design and implement this integration using a HDF5 Virtual File Driver (VFD). The GDS VFD provides a file system abstraction to the application that allows HDF5 applications to perform I/O without the need to move data between CPUs and GPUs explicitly. We compare performance of the HDF5 GDS VFD with explicit data movement approaches and demonstrate superior performance with the GDS method.
- Research Organization:
- Lawrence Berkeley National Laboratory (LBNL), Berkeley, CA (United States)
- Sponsoring Organization:
- USDOE Office of Science (SC)
- DOE Contract Number:
- AC02-05CH11231
- OSTI ID:
- 1760260
- Country of Publication:
- United States
- Language:
- English
Similar Records
HDF5 in the exascale era: Delivering efficient and scalable parallel I/O for exascale applications
Tuning HDF5 subfiling performance on parallel file systems
Tuning HDF5 for Lustre File Systems
Journal Article
·
Tue Oct 15 20:00:00 EDT 2024
· International Journal of High Performance Computing Applications
·
OSTI ID:2573001
Tuning HDF5 subfiling performance on parallel file systems
Conference
·
Fri May 12 00:00:00 EDT 2017
·
OSTI ID:1398484
Tuning HDF5 for Lustre File Systems
Conference
·
Fri Sep 24 00:00:00 EDT 2010
·
OSTI ID:1050648