Validating the simulation of large-scale parallel applications using statistical characteristics
Abstract
Simulation is a widely adopted method to analyze and predict the performance of large-scale parallel applications. Validating the hardware model is highly important for complex simulations with a large number of parameters. Common practice involves calculating the percent error between the projected and the real execution time of a benchmark program. However, in a high-dimensional parameter space, this coarse-grained approach often suffers from parameter insensitivity, which may not be known a priori. Moreover, the traditional approach cannot be applied to the validation of software models, such as application skeletons used in online simulations. In this work, we present a methodology and a toolset for validating both hardware and software models by quantitatively comparing fine-grained statistical characteristics obtained from execution traces. Although statistical information has been used in tasks like performance optimization, this is the first attempt to apply it to simulation validation. Lastly, our experimental results show that the proposed evaluation approach offers significant improvement in fidelity when compared to evaluation using total execution time, and the proposed metrics serve as reliable criteria that progress toward automating the simulation tuning process.
- Authors:
-
- Univ. of Central Florida, Orlando, FL (United States)
- Sandia National Lab. (SNL-CA), Livermore, CA (United States)
- Univ. of Central Florida, Orlando, FL (United States); Sandia National Lab. (SNL-CA), Livermore, CA (United States)
- Publication Date:
- Research Org.:
- Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Sandia National Lab. (SNL-CA), Livermore, CA (United States)
- Sponsoring Org.:
- USDOE National Nuclear Security Administration (NNSA)
- OSTI Identifier:
- 1333867
- Report Number(s):
- SAND-2015-2905J
Journal ID: ISSN 2376-3639; 583296
- Grant/Contract Number:
- AC04-94AL85000
- Resource Type:
- Accepted Manuscript
- Journal Name:
- ACM Transactions on Modeling and Performance Evaluation of Computing Systems
- Additional Journal Information:
- Journal Volume: 1; Journal Issue: 1; Journal ID: ISSN 2376-3639
- Publisher:
- Association for Computing Machinery
- Country of Publication:
- United States
- Language:
- English
- Subject:
- 97 MATHEMATICS AND COMPUTING; measurement; experimentation; simulation evaluation; evaluation metrics; software skeleton
Citation Formats
Zhang, Deli, Wilke, Jeremiah, Hendry, Gilbert, and Dechev, Damian. Validating the simulation of large-scale parallel applications using statistical characteristics. United States: N. p., 2016.
Web. doi:10.1145/2809778.
Zhang, Deli, Wilke, Jeremiah, Hendry, Gilbert, & Dechev, Damian. Validating the simulation of large-scale parallel applications using statistical characteristics. United States. https://doi.org/10.1145/2809778
Zhang, Deli, Wilke, Jeremiah, Hendry, Gilbert, and Dechev, Damian. Tue .
"Validating the simulation of large-scale parallel applications using statistical characteristics". United States. https://doi.org/10.1145/2809778. https://www.osti.gov/servlets/purl/1333867.
@article{osti_1333867,
title = {Validating the simulation of large-scale parallel applications using statistical characteristics},
author = {Zhang, Deli and Wilke, Jeremiah and Hendry, Gilbert and Dechev, Damian},
abstractNote = {Simulation is a widely adopted method to analyze and predict the performance of large-scale parallel applications. Validating the hardware model is highly important for complex simulations with a large number of parameters. Common practice involves calculating the percent error between the projected and the real execution time of a benchmark program. However, in a high-dimensional parameter space, this coarse-grained approach often suffers from parameter insensitivity, which may not be known a priori. Moreover, the traditional approach cannot be applied to the validation of software models, such as application skeletons used in online simulations. In this work, we present a methodology and a toolset for validating both hardware and software models by quantitatively comparing fine-grained statistical characteristics obtained from execution traces. Although statistical information has been used in tasks like performance optimization, this is the first attempt to apply it to simulation validation. Lastly, our experimental results show that the proposed evaluation approach offers significant improvement in fidelity when compared to evaluation using total execution time, and the proposed metrics serve as reliable criteria that progress toward automating the simulation tuning process.},
doi = {10.1145/2809778},
journal = {ACM Transactions on Modeling and Performance Evaluation of Computing Systems},
number = 1,
volume = 1,
place = {United States},
year = {Tue Mar 01 00:00:00 EST 2016},
month = {Tue Mar 01 00:00:00 EST 2016}
}