Skip to main content
U.S. Department of Energy
Office of Scientific and Technical Information

A Pipeline for Large Data Processing Using Regular Sampling for Unstructured Grids

Technical Report ·
DOI:https://doi.org/10.2172/1357102· OSTI ID:1357102
 [1];  [2];  [3];  [2];  [1]
  1. Los Alamos National Lab. (LANL), Los Alamos, NM (United States)
  2. Virginia Polytechnic Inst. and State Univ. (Virginia Tech), Blacksburg, VA (United States)
  3. Univ. of Texas, Austin, TX (United States)
Large simulation data requires a lot of time and computational resources to compute, store, analyze, visualize, and run user studies. Today, the largest cost of a supercomputer is not hardware but maintenance, in particular energy consumption. Our goal is to balance energy consumption and cognitive value of visualizations of resulting data. This requires us to go through the entire processing pipeline, from simulation to user studies. To reduce the amount of resources, data can be sampled or compressed. While this adds more computation time, the computational overhead is negligible compared to the simulation time. We built a processing pipeline at the example of regular sampling. The reasons for this choice are two-fold: using a simple example reduces unnecessary complexity as we know what to expect from the results. Furthermore, it provides a good baseline for future, more elaborate sampling methods. We measured time and energy for each test we did, and we conducted user studies in Amazon Mechanical Turk (AMT) for a range of different results we produced through sampling.
Research Organization:
Los Alamos National Lab. (LANL), Los Alamos, NM (United States)
Sponsoring Organization:
USDOE Office of Science (SC)
DOE Contract Number:
AC52-06NA25396
OSTI ID:
1357102
Report Number(s):
LA--UR-17-23903
Country of Publication:
United States
Language:
English