skip to main content
OSTI.GOV title logo U.S. Department of Energy
Office of Scientific and Technical Information

Title: Enabling communication concurrency through flexible MPI endpoints

Journal Article · · International Journal of High Performance Computing Applications
 [1];  [2];  [3];  [4];  [5];  [3];  [3]
  1. Intel Corporation, Hudson, MA (United States)
  2. Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)
  3. Argonne National Lab. (ANL), Argonne, IL (United States)
  4. Cisco Systems Inc., San Jose, CA (United States)
  5. International Business Machines Corporation, Rochester, MN (United States)

MPI defines a one-to-one relationship between MPI processes and ranks. This model captures many use cases effectively; however, it also limits communication concurrency and interoperability between MPI and programming models that utilize threads. Our paper describes the MPI endpoints extension, which relaxes the longstanding one-to-one relationship between MPI processes and ranks. Using endpoints, an MPI implementation can map separate communication contexts to threads, allowing them to drive communication independently. Also, endpoints enable threads to be addressable in MPI operations, enhancing interoperability between MPI and other programming models. Furthermore, these characteristics are illustrated through several examples and an empirical study that contrasts current multithreaded communication performance with the need for high degrees of communication concurrency to achieve peak communication performance.

Research Organization:
Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)
Sponsoring Organization:
USDOE National Nuclear Security Administration (NNSA)
Grant/Contract Number:
AC04-94AL85000
OSTI ID:
1140752
Report Number(s):
SAND2014-0614J; 498469
Journal Information:
International Journal of High Performance Computing Applications, Vol. 28, Issue 4; ISSN 1094-3420
Publisher:
SAGECopyright Statement
Country of Publication:
United States
Language:
English
Citation Metrics:
Cited by: 30 works
Citation information provided by
Web of Science

References (23)

The impact of hybrid-core processors on MPI message rate conference January 2013
Hybrid PGAS runtime support for multicore nodes conference January 2010
Generalized communicators in the message passing interface journal June 2001
Enabling MPI interoperability through flexible communication endpoints conference January 2013
Hybrid parallel programming with MPI and unified parallel C conference January 2010
Enabling Concurrent Multithreaded MPI Communication on Multicore Petascale Systems book January 2010
Ownership passing: efficient distributed memory programming on multi-core systems
  • Friedley, Andrew; Hoefler, Torsten; Bronevetsky, Greg
  • Proceedings of the 18th ACM SIGPLAN symposium on Principles and practice of parallel programming - PPoPP '13 https://doi.org/10.1145/2442516.2442534
conference January 2013
Dynamic Communicators in MPI book January 2009
MPI + MPI: a new hybrid approach to parallel programming with MPI plus shared memory journal May 2013
Supporting Hybrid MPI and OpenSHMEM over InfiniBand: Design and Performance Evaluation conference September 2012
Unifying UPC and MPI runtimes: experience with MVAPICH conference January 2010
FG-MPI: Fine-grain MPI for multicore and clusters
  • Kamal, Humaira; Wagner, Alan
  • Distributed Processing, Workshops and Phd Forum (IPDPSW 2010), 2010 IEEE International Symposium on Parallel & Distributed Processing, Workshops and Phd Forum (IPDPSW) https://doi.org/10.1109/IPDPSW.2010.5470773
conference April 2010
PAMI: A Parallel Active Message Interface for the Blue Gene/Q Supercomputer
  • Kumar, Sameer; Mamidala, Amith R.; Faraj, Daniel A.
  • 2012 IEEE International Symposium on Parallel & Distributed Processing (IPDPS), 2012 IEEE 26th International Parallel and Distributed Processing Symposium https://doi.org/10.1109/IPDPS.2012.73
conference May 2012
NUMA-aware shared-memory collective communication for MPI
  • Li, Shigang; Hoefler, Torsten; Snir, Marc
  • Proceedings of the 22nd international symposium on High-performance parallel and distributed computing - HPDC '13 https://doi.org/10.1145/2493123.2462903
conference January 2013
Multi-threaded UPC runtime with network endpoints: Design alternatives and evaluation on multi-core architectures conference December 2011
Hybrid MPI/OpenMP Parallel Programming on Clusters of Multi-Core SMP Nodes
  • Rabenseifner, Rolf; Hager, Georg; Jost, Gabriele
  • 2009 17th Euromicro International Conference on Parallel, Distributed and Network-based Processing https://doi.org/10.1109/PDP.2009.43
conference February 2009
Development of Mixed Mode MPI / OpenMP Applications journal January 2001
Extending MPI to accelerators conference January 2011
Network Endpoints for Clusters of SMPs
  • Tanase, Gabriel; Almasi, Gheorghe; Xue, Hanhong
  • 2012 24th International Symposium on Computer Architecture and High Performance Computing (SBAC-PAD), 2012 IEEE 24th International Symposium on Computer Architecture and High Performance Computing https://doi.org/10.1109/SBAC-PAD.2012.15
conference October 2012
Compact and Efficient Implementation of the MPI Group Operations book January 2010
Evaluating NIC hardware requirements to achieve high message rate PGAS support on multi-core processors conference January 2007
MVAPICH2-GPU: optimized GPU to GPU communication for InfiniBand clusters journal April 2011
Portable, MPI-interoperable coarray fortran
  • Yang, Chaoran; Bland, Wesley; Mellor-Crummey, John
  • Proceedings of the 19th ACM SIGPLAN symposium on Principles and practice of parallel programming - PPoPP '14 https://doi.org/10.1145/2555243.2555270
conference January 2014

Cited By (6)

Optimizing point‐to‐point communication between adaptive MPI endpoints in shared memory journal March 2018
A survey of MPI usage in the US exascale computing project: A survey of MPI usage in the U. S. exascale computing project journal September 2018
Hardware MPI message matching: Insights into MPI matching behavior to inform design
  • Ferreira, Kurt; Grant, Ryan E.; Levenhagen, Michael J.
  • Concurrency and Computation: Practice and Experience, Vol. 32, Issue 3 https://doi.org/10.1002/cpe.5150
journal February 2019
Tail queues: A multi‐threaded matching architecture
  • Dosanjh, Matthew G. F.; Grant, Ryan E.; Schonbein, Whit
  • Concurrency and Computation: Practice and Experience, Vol. 32, Issue 3 https://doi.org/10.1002/cpe.5158
journal February 2019
Finepoints: Partitioned Multithreaded MPI Communication
  • Grant, Ryan E.; Dosanjh, Matthew G. F.; Levenhagen, Michael J.
  • High Performance Computing: 34th International Conference, ISC High Performance 2019, Frankfurt/Main, Germany, June 16–20, 2019, Proceedings, p. 330-350 https://doi.org/10.1007/978-3-030-20656-7_17
book May 2019
MPI Sessions: Leveraging Runtime Infrastructure to Increase Scalability of Applications at Exascale conference January 2016

Similar Records

Enabling communication concurrency through flexible MPI endpoints
Journal Article · Tue Sep 23 00:00:00 EDT 2014 · International Journal of High Performance Computing Applications · OSTI ID:1140752

Optimizing point‐to‐point communication between adaptive MPI endpoints in shared memory
Journal Article · Mon Mar 12 00:00:00 EDT 2018 · Concurrency and Computation. Practice and Experience · OSTI ID:1140752

Test suite for evaluating performance of multithreaded MPI communication.
Journal Article · Tue Dec 01 00:00:00 EST 2009 · Parallel Comput. · OSTI ID:1140752