skip to main content
OSTI.GOV title logo U.S. Department of Energy
Office of Scientific and Technical Information

Title: A Programming Model for Massive Data Parallelism with Data Dependencies

Abstract

Accelerating processors can often be more cost and energy effective for a wide range of data-parallel computing problems than general-purpose processors. For graphics processor units (GPUs), this is particularly the case when program development is aided by environments such as NVIDIA s Compute Unified Device Architecture (CUDA), which dramatically reduces the gap between domain-specific architectures and general purpose programming. Nonetheless, general-purpose GPU (GPGPU) programming remains subject to several restrictions. Most significantly, the separation of host (CPU) and accelerator (GPU) address spaces requires explicit management of GPU memory resources, especially for massive data parallelism that well exceeds the memory capacity of GPUs. One solution to this problem is to transfer data between the GPU and host memories frequently. In this work, we investigate another approach. We run massively data-parallel applications on GPU clusters. We further propose a programming model for massive data parallelism with data dependencies for this scenario. Experience from micro benchmarks and real-world applications shows that our model provides not only ease of programming but also significant performance gains.

Authors:
 [1];  [2];  [1];  [1]
  1. ORNL
  2. North Carolina State University
Publication Date:
Research Org.:
Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States)
Sponsoring Org.:
USDOE Laboratory Directed Research and Development (LDRD) Program; Work for Others (WFO)
OSTI Identifier:
964332
DOE Contract Number:  
DE-AC05-00OR22725
Resource Type:
Conference
Resource Relation:
Conference: Parallel Architectures and Compilation Techniques (PACT), Raleigh, NC, USA, 20090912, 20090912
Country of Publication:
United States
Language:
English
Subject:
43 PARTICLE ACCELERATORS; ACCELERATORS; ARCHITECTURE; BENCHMARKS; CAPACITY; MANAGEMENT; PERFORMANCE; PROGRAMMING

Citation Formats

Cui, Xiaohui, Mueller, Frank, Potok, Thomas E, and Zhang, Yongpeng. A Programming Model for Massive Data Parallelism with Data Dependencies. United States: N. p., 2009. Web.
Cui, Xiaohui, Mueller, Frank, Potok, Thomas E, & Zhang, Yongpeng. A Programming Model for Massive Data Parallelism with Data Dependencies. United States.
Cui, Xiaohui, Mueller, Frank, Potok, Thomas E, and Zhang, Yongpeng. Thu . "A Programming Model for Massive Data Parallelism with Data Dependencies". United States. doi:.
@article{osti_964332,
title = {A Programming Model for Massive Data Parallelism with Data Dependencies},
author = {Cui, Xiaohui and Mueller, Frank and Potok, Thomas E and Zhang, Yongpeng},
abstractNote = {Accelerating processors can often be more cost and energy effective for a wide range of data-parallel computing problems than general-purpose processors. For graphics processor units (GPUs), this is particularly the case when program development is aided by environments such as NVIDIA s Compute Unified Device Architecture (CUDA), which dramatically reduces the gap between domain-specific architectures and general purpose programming. Nonetheless, general-purpose GPU (GPGPU) programming remains subject to several restrictions. Most significantly, the separation of host (CPU) and accelerator (GPU) address spaces requires explicit management of GPU memory resources, especially for massive data parallelism that well exceeds the memory capacity of GPUs. One solution to this problem is to transfer data between the GPU and host memories frequently. In this work, we investigate another approach. We run massively data-parallel applications on GPU clusters. We further propose a programming model for massive data parallelism with data dependencies for this scenario. Experience from micro benchmarks and real-world applications shows that our model provides not only ease of programming but also significant performance gains.},
doi = {},
journal = {},
number = ,
volume = ,
place = {United States},
year = {Thu Jan 01 00:00:00 EST 2009},
month = {Thu Jan 01 00:00:00 EST 2009}
}

Conference:
Other availability
Please see Document Availability for additional information on obtaining the full-text document. Library patrons may search WorldCat to identify libraries that hold this conference proceeding.

Save / Share: