skip to main content
OSTI.GOV title logo U.S. Department of Energy
Office of Scientific and Technical Information

Title: Automatic Parallelization and Transparent Fault Tolerance

Authors:
 [1];  [1];  [1];  [2];  [3]
  1. Los Alamos National Laboratory
  2. University of North Dakota
  3. New Mexico Tech
Publication Date:
Research Org.:
Los Alamos National Lab. (LANL), Los Alamos, NM (United States)
Sponsoring Org.:
USDOE Laboratory Directed Research and Development (LDRD) Program
OSTI Identifier:
1258367
Report Number(s):
LA-UR-16-24056
DOE Contract Number:
AC52-06NA25396
Resource Type:
Conference
Resource Relation:
Conference: 17th Symposium on Trends in Functional Programming ; 2016-06-08 - 2016-06-10 ; College Park, Maryland, United States
Country of Publication:
United States
Language:
English
Subject:
Computer Science

Citation Formats

Davis, Marion Kei, Prichard, Dean A., Ringo, David Matteson, Anderson, Loren, and Marks, Jacob. Automatic Parallelization and Transparent Fault Tolerance. United States: N. p., 2016. Web.
Davis, Marion Kei, Prichard, Dean A., Ringo, David Matteson, Anderson, Loren, & Marks, Jacob. Automatic Parallelization and Transparent Fault Tolerance. United States.
Davis, Marion Kei, Prichard, Dean A., Ringo, David Matteson, Anderson, Loren, and Marks, Jacob. 2016. "Automatic Parallelization and Transparent Fault Tolerance". United States. doi:. https://www.osti.gov/servlets/purl/1258367.
@article{osti_1258367,
title = {Automatic Parallelization and Transparent Fault Tolerance},
author = {Davis, Marion Kei and Prichard, Dean A. and Ringo, David Matteson and Anderson, Loren and Marks, Jacob},
abstractNote = {},
doi = {},
journal = {},
number = ,
volume = ,
place = {United States},
year = 2016,
month = 6
}

Conference:
Other availability
Please see Document Availability for additional information on obtaining the full-text document. Library patrons may search WorldCat to identify libraries that hold this conference proceeding.

Save / Share:
  • Checkpoint/restart (C/R) has become a requirement for long-running jobs in large-scale clusters due to a meantime- to-failure (MTTF) in the order of hours. After a failure, C/R mechanisms generally require a complete restart of an MPI job from the last checkpoint. A complete restart, however, is unnecessary since all but one node are typically still alive. Furthermore, a restart may result in lengthy job requeuing even though the original job had not exceeded its time quantum. In this paper, we overcome these shortcomings. Instead of job restart, we have developed a transparent mechanism for job pause within LAM/MPI+BLCR. This mechanismmore » allows live nodes to remain active and roll back to the last checkpoint while failed nodes are dynamically replaced by spares before resuming from the last checkpoint. Our methodology includes LAM/MPI enhancements in support of scalable group communicationwith fluctuating number of nodes, reuse of network connections, transparent coordinated checkpoint scheduling and a BLCR enhancement for job pause. Experiments in a cluster with the NAS Parallel Benchmark suite show that our overhead for job pause is comparable to that of a complete job restart. A minimal overhead of 5.6% is only incurred in case migration takes place while the regular checkpoint overhead remains unchanged. Yet, our approach alleviates the need to reboot the LAM run-time environment, which accounts for considerable overhead resulting in net savings of our scheme in the experiments. Our solution further provides full transparency and automation with the additional benefit of reusing existing resources. Executing continues after failures within the scheduled job, i.e., the application staging overhead is not incurred again in contrast to a restart. Our scheme offers additional potential for savings through incremental checkpointing and proactive diskless live migration, which we are currently working on.« less
  • As the core count of HPC machines continue to grow in size, issues such as fault tolerance and reliability are becoming limiting factors for application scalability. Current techniques to ensure progress across faults, for example coordinated checkpoint-restart, are unsuitable for machines of this scale due to their predicted high overheads. In this study, we present the design and implementation of a novel system for ensuring reliability which uses transparent, rank-level, redundant computation. Using this system, we show the overheads involved in redundant computation for a number of real-world HPC applications. Additionally, we relate the communication characteristics of an application tomore » the overheads observed.« less
  • Detailed attempts at realistic nuclear reactor simulations currently take many times real time to execute on high performance workstations. Even the fastest sequential machine can not run these simulations fast enough to ensure that the best corrective measure is used during a nuclear accident to prevent a minor malfunction from becoming a major catastrophe. Since sequential computers have nearly reached the speed of light barrier, these simulations will have to be run in parallel to make significant improvements in speed. In physical reactor plants, parallelism abounds. Fluids flow, controls change, and reactions occur in parallel with only adjacent components directlymore » affecting each other. These do not occur in the sequentialized manner, with global instantaneous effects, that is often used in simulators. Development of parallel algorithms that more closely approximate the real-world operation of a reactor may, in addition to speeding up the simulations, actually improve the accuracy and reliability of the predictions generated. Three types of parallel architecture (shared memory machines, distributed memory multicomputers, and distributed networks) are briefly reviewed as targets for parallelization of nuclear reactor simulation. Various parallelization models (loop-based model, shared memory model, functional model, data parallel model, and a combined functional and data parallel model) are discussed along with their advantages and disadvantages for nuclear reactor simulation. A variety of tools are introduced for each of the models. Emphasis is placed on the data parallel model as the primary focus for two-phase flow simulation. Tools to support data parallel programming for multiple component applications and special parallelization considerations are also discussed.« less
  • A Straight-line code, which consists of assignment, addition, and multiplication statements is an abstraction of a serial computer program to compute a function with n inputs. Given a serial straight-line code with N statements, the authors derive an algorithm that automatically evaluates not only the function but also its first-order derivatives with respect to the n inputs on a parallel computer. The basic idea of the algorithm is to marry automatic computation of derivatives with automatic parallelization of serial programs. The algorithm requires O(M{sub N} log of N) scalar operations, where O(M{sub N}) is the time complexity of a parallelmore » multiplication of two dense N x N matrices and it represents a measure of the complexity of the straight-line code. Although it can be exponential in N in the worse case, it tends to be only polynomial in N for many important problems.« less