Skip to main content
U.S. Department of Energy
Office of Scientific and Technical Information

Exploiting Lustre File Joining for Effective Collective IO

Conference ·
OSTI ID:931507
Lustre is a parallel file system that presents high aggregated IO bandwidth by striping file extents across many storage devices. However, our experiments indicate excessively wide striping can cause performance degradation. Lustre supports an innovative file joining feature that joins files in place. To mitigate striping overhead and benefit collective IO, we propose two techniques: split writing and hierarchical striping. In split writing, a file is created as separate subfiles, each of which is striped to only a few storage devices. They are joined as a single file at the file close time. Hierarchical striping builds on top of split writing and orchestrates the span of subfiles in a hierarchical manner to avoid overlapping and achieve the appropriate coverage of storage devices. Together, these techniques can avoid the overhead associated with large stripe width, while still being able to combine bandwidth available from many storage devices. We have prototyped these techniques in the ROMIO implementation of MPI-IO. Experimental results indicate that split writing and hierarchical striping can significantly improve the performance of Lustre collective IO in terms of both data transfer and management operations. On a Lustre file system configured with 46 object storage targets, our implementation improves collective write performance of a 16-process job by as much as 220%.
Research Organization:
Oak Ridge National Laboratory (ORNL); Center for Computational Sciences
Sponsoring Organization:
SC USDOE - Office of Science (SC)
DOE Contract Number:
AC05-00OR22725
OSTI ID:
931507
Country of Publication:
United States
Language:
English

Similar Records

Making resonance a common case: a high-performance implementation of collective I/O on parallel file systems
Conference · Wed Dec 31 23:00:00 EST 2008 · OSTI ID:956499

Tuning HDF5 subfiling performance on parallel file systems
Conference · Fri May 12 00:00:00 EDT 2017 · OSTI ID:1398484

Implementing MPI-IO atomic mode and shared file pointers using MPI one-sided communication.
Journal Article · Sun Jul 01 00:00:00 EDT 2007 · Int. J. High Perform. Comput. Appl. · OSTI ID:914949