skip to main content
OSTI.GOV title logo U.S. Department of Energy
Office of Scientific and Technical Information

Title: Tuning collective communication for Partitioned Global Address Space programming models

Journal Article · · Parallel Computing
 [1];  [2];  [2];  [2]
  1. Univ. of California, Berkeley, CA (United States)
  2. Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States)

Partitioned Global Address Space (PGAS) languages offer programmers the convenience of a shared memory programming style combined with locality control necessary to run on large-scale distributed memory systems. Even within a PGAS language programmers often need to perform global communication operations such as broadcasts or reductions, which are best performed as collective operations in which a group of threads work together to perform the operation. In this study we consider the problem of implementing collective communication within PGAS languages and explore some of the design trade-offs in both the interface and implementation. In particular, PGAS collectives have semantic issues that are different than in send–receive style message passing programs, and different implementation approaches that take advantage of the one-sided communication style in these languages. We present an implementation framework for PGAS collectives as part of the GASNet communication layer, which supports shared memory, distributed memory and hybrids. The framework supports a broad set of algorithms for each collective, over which the implementation may be automatically tuned. In conclusion, we demonstrate the benefit of optimized GASNet collectives using application benchmarks written in UPC, and demonstrate that the GASNet collectives can deliver scalable performance on a variety of state-of-the-art parallel machines including a Cray XT4, an IBM BlueGene/P, and a Sun Constellation system with InfiniBand interconnect.

Research Organization:
Lawrence Berkeley National Laboratory (LBNL), Berkeley, CA (United States)
Sponsoring Organization:
USDOE Office of Science (SC), Advanced Scientific Computing Research (ASCR); National Science Foundation (NSF)
Grant/Contract Number:
AC02-05CH11231; FC03-01ER25509; FC02-07ER25799; AC02-06CH11357; AC05-00OR22725; OCI-0749190
OSTI ID:
1407106
Journal Information:
Parallel Computing, Vol. 37, Issue 9; ISSN 0167-8191
Publisher:
ElsevierCopyright Statement
Country of Publication:
United States
Language:
English
Citation Metrics:
Cited by: 22 works
Citation information provided by
Web of Science

References (7)

Co-array Fortran for parallel programming journal August 1998
Parallel Programmability and the Chapel Language journal August 2007
Implications of application usage characteristics for collective communication offload journal January 2006
Efficient algorithms for all-to-all communications in multiport message-passing systems journal January 1997
Optimizing bandwidth limited problems using one-sided communication and overlap conference January 2006
The Nas Parallel Benchmarks journal September 1991
Scaling communication-intensive applications on BlueGene/P using one-sided communication and overlap conference May 2009

Cited By (4)

A view of programming scalable data analysis: from clouds to exascale journal February 2019
GASNet-EX: A High-Performance, Portable Communication Library for Exascale
  • Bonachea, Dan; Hargrove, Paul H.; Hall, Mary
  • Languages and Compilers for Parallel Computing: 31st International Workshop, LCPC 2018, Salt Lake City, UT, USA, October 9–11, 2018, Revised Selected Papers, p. 138-158 https://doi.org/10.1007/978-3-030-34627-0_11
book November 2019
GASNet-EX: A High-Performance, Portable Communication Library for Exascale report October 2018
Parallel and scalable short-read alignment on multi-core clusters using UPC++ text January 2016