Skip to main content
U.S. Department of Energy
Office of Scientific and Technical Information

Work stealing for GPU-accelerated parallel programs in a global address space framework

Journal Article · · Concurrency and Computation. Practice and Experience
DOI:https://doi.org/10.1002/cpe.3747· OSTI ID:1393474
 [1];  [2];  [3];  [2];  [1]
  1. Department of Computer Science and Engineering, The Ohio State University, Columbus OH USA
  2. Mathematics and Computer Science Division, Argonne National Laboratory, Lemont IL USA
  3. Computer Science and Mathematics Division, Pacific Northwest National Laboratory, Richland WA USA
Task parallelism is an attractive approach to automatically load balance the computation in a parallel system and adapt to dynamism exhibited by parallel systems. Exploiting task parallelism through work stealing has been extensively studied in shared and distributed-memory contexts. In this paper, we study the design of a system that uses work stealing for dynamic load balancing of task-parallel programs executed on hybrid distributed-memory CPU-graphics processing unit (GPU) systems in a global-address space framework. We take into account the unique nature of the accelerator model employed by GPUs, the significant performance difference between GPU and CPU execution as a function of problem size, and the distinct CPU and GPU memory domains. We consider various alternatives in designing a distributed work stealing algorithm for CPU-GPU systems, while taking into account the impact of task distribution and data movement overheads. These strategies are evaluated using microbenchmarks that capture various execution configurations as well as the state-of-the-art CCSD(T) application module from the computational chemistry domain
Research Organization:
Argonne National Laboratory (ANL)
Sponsoring Organization:
USDOE Office of Science
DOE Contract Number:
AC02-06CH11357
OSTI ID:
1393474
Journal Information:
Concurrency and Computation. Practice and Experience, Journal Name: Concurrency and Computation. Practice and Experience Journal Issue: 13 Vol. 28; ISSN 1532-0626
Publisher:
Wiley
Country of Publication:
United States
Language:
English

References (13)

Advances, Applications and Performance of the Global Arrays Shared Memory Programming Toolkit journal May 2006
ScalaBLAST: A Scalable Implementation of BLAST for High-Performance Data-Intensive Bioinformatics Analysis journal August 2006
Co-array Fortran for parallel programming journal August 1998
Designing Reliable Systems from Unreliable Components: The Challenges of Transistor Variability and Degradation journal November 2005
High performance computational chemistry: An overview of NWChem a distributed parallel application journal June 2000
Scalable Load Balancing Techniques for Parallel Computers journal July 1994
Coupled-cluster theory in quantum chemistry journal February 2007
Towards dense linear algebra for hybrid GPU accelerated manycore systems journal June 2010
Synthesis of High-Performance Parallel Programs for a Class of ab Initio Quantum Chemistry Models journal February 2005
Lifeline-based global load balancing journal February 2011
Titanium: a high-performance Java dialect journal September 1998
Parallel Programmability and the Chapel Language journal August 2007
Bounds on Multiprocessing Timing Anomalies journal March 1969

Similar Records

Work stealing for GPU-accelerated parallel programs in a global address space framework: WORK STEALING ON GPU-ACCELERATED SYSTEMS
Journal Article · Tue Jan 05 23:00:00 EST 2016 · Concurrency and Computation. Practice and Experience · OSTI ID:1333989

Scalable Work Stealing
Conference · Fri Nov 13 23:00:00 EST 2009 · OSTI ID:986715

Data-driven Fault Tolerance for Work Stealing Computations
Conference · Mon Jun 25 00:00:00 EDT 2012 · OSTI ID:1239507