skip to main content
OSTI.GOV title logo U.S. Department of Energy
Office of Scientific and Technical Information

Title: Experiences in extending parallware to support OpenACC. In: WACCPD '15 Proceedings of the Second Workshop on Accelerator Programming using Directives, Article No. 4

Conference ·
 [1];  [2];  [3]
  1. Appentra Solutions, A Coruna (Spain)
  2. Appentra Solutions, A Coruna (Spain); Univ. of A Coruna (Spain)
  3. Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States)

Porting scientific codes to accelerator-based computers using OpenACC and OpenMP is an important topic for the HPC community. Programmability, performance portability and developer productivity are key issues for the widespread use of these systems. In the scope of general-purpose parallel computing, Parallware is a new commercial OpenMP-enabling source-to-source compiler that automatically adds OpenMP capabilities in scientific programs. Thus, extending Parallware with OpenACC or OpenMP 4.x support would contribute to improve programmability and developer productivity. In contrast, the performance portability of such approach needs to be demonstrated in practice. This paper presents a preliminary study to extend Parallware with OpenACC support for GPU devices. A simple benchmark suite has been designed to mimic important features and computational patterns of real scientific applications. Handcoded OpenACC versions are compared to OpenMP versions automatically generated by Parallware. Performance is evaluated with the PGI OpenACC compiler on systems accelerated with NVIDIA GPUs.

Research Organization:
Oak Ridge National Laboratory (ORNL), Oak Ridge, TN (United States). Oak Ridge Leadership Computing Facility (OLCF)
Sponsoring Organization:
USDOE Office of Science (SC)
OSTI ID:
1567642
Resource Relation:
Conference: Second Workshop on Accelerator Programming using Directives Austin, Texas — November 15 - 15, 2015
Country of Publication:
United States
Language:
English

References (15)

A novel compiler support for automatic parallelization on multicore systems journal September 2013
A GSA-based compiler infrastructure to extract parallelism from complex loops conference January 2003
The NAS parallel benchmarks---summary and preliminary results conference January 1991
Parallel programming with Polaris journal January 1996
Anatomy of high-performance matrix multiplication journal May 2008
Auto-tuning a high-level language targeted to GPU codes conference May 2012
Optimizing Scientific Workflows in the Cloud: A Montage Example conference December 2014
The ParaWise Expert Assistant – Widening Accessibility to Efficient and Scalable Tool Generated OpenMP Code book January 2005
A unified framework for optimizing locality, parallelism, and communication in out-of-core computations journal July 2000
SUIF Explorer: an interactive and interprocedural parallelizer
  • Liao, Shih-Wei; Diwan, Amer; Bosch, Robert P.
  • Proceedings of the seventh ACM SIGPLAN symposium on Principles and practice of parallel programming - PPoPP '99 https://doi.org/10.1145/301104.301108
conference January 1999
Exploiting locality in the run-time parallelization of irregular loops conference January 2002
A compiler optimization algorithm for shared-memory multiprocessors journal January 1998
Optimal weighted loop fusion for parallel programs conference January 1997
Combining loop transformations considering caches and scheduling conference January 1996
Adaptive reduction parallelization techniques conference January 2000

Similar Records

Related Subjects