Skip to main content
U.S. Department of Energy
Office of Scientific and Technical Information

Exploiting hierarchy in parallel computer networks to optimize collective operations performance

Conference ·

The efficient implementation of collective communication operations has received much attention. Initial efforts modeled network communication and produced optimal trees based on those models. However, the models used by these initial efforts assumed equal point-to-point latencies between any two processes. This assumption is violated in heterogeneous systems such as clusters of SMPs and wide-area computational grids, and as a result, collective operations that utilize the trees generated by these models perform suboptimally. In response, more recent work has focused on creating topology-aware trees for collective operations that minimize communication across slower channels (e.g., a wide-area network). While these efforts have significant communication benefits, they all limit their view of the network to only two layers. The authors present a strategy based upon a multilayer view of the network. By creating multilevel topology trees they take advantage of communication cost differences at every level in the network. They used this strategy to implement topology-aware versions of several MPI collective operations in MPICH-G, the Globus-enabled version of the popular MPICH implementation of the MPI standard. Using information about topology discovered by Globus, they construct these topology-aware trees automatically during execution, thus freeing the MPI application programmer from having to write special files or functions to describe the topology to the MPICH library. They present results demonstrating the advantages of their multilevel approach by comparing it to the default (topology-unaware) implementation provided by MPICH and a topology-aware two-layer implementation.

Research Organization:
Argonne National Lab., IL (US)
Sponsoring Organization:
US Department of Energy (US)
DOE Contract Number:
W-31109-ENG-38
OSTI ID:
764215
Report Number(s):
ANL/MCS/CP-101052
Country of Publication:
United States
Language:
English

Similar Records

Technologies and tools for high-performance distributed computing. Final report
Technical Report · Mon May 01 00:00:00 EDT 2000 · OSTI ID:808724

MPICH-G2 : a grid-enabled implementation of the message passing interface.
Journal Article · Thu May 01 00:00:00 EDT 2003 · J. Parallel Distrib. Comput. · OSTI ID:949654

Wide-area implementation of the message passing interface.
Journal Article · Wed Dec 31 23:00:00 EST 1997 · Parallel Comput. · OSTI ID:943164