skip to main content
DOE PAGES title logo U.S. Department of Energy
Office of Scientific and Technical Information

Title: CMS Analysis and Data Reduction with Apache Spark

Abstract

Experimental Particle Physics has been at the forefront of analyzing the world's largest datasets for decades. The HEP community was among the first to develop suitable software and computing tools for this task. In recent times, new toolkits and systems for distributed data processing, collectively called "Big Data" technologies have emerged from industry and open source projects to support the analysis of Petabyte and Exabyte datasets in industry. While the principles of data analysis in HEP have not changed (filtering and transforming experiment-specific data formats), these new technologies use different approaches and tools, promising a fresh look at analysis of very large datasets that could potentially reduce the time-to-physics with increased interactivity. Moreover these new tools are typically actively developed by large communities, often profiting of industry resources, and under open source licensing. These factors result in a boost for adoption and maturity of the tools and for the communities supporting them, at the same time helping in reducing the cost of ownership for the end-users. In this talk, we are presenting studies of using Apache Spark for end user data analysis. We are studying the HEP analysis workflow separated into two thrusts: the reduction of centrally produced experiment datasetsmore » and the end-analysis up to the publication plot. Studying the first thrust, CMS is working together with CERN openlab and Intel on the CMS Big Data Reduction Facility. The goal is to reduce 1 PB of official CMS data to 1 TB of ntuple output for analysis. We are presenting the progress of this 2-year project with first results of scaling up Spark-based HEP analysis. Studying the second thrust, we are presenting studies on using Apache Spark for a CMS Dark Matter physics search, comparing Spark's feasibility, usability and performance to the ROOT-based analysis.« less

Authors:
ORCiD logo [1];  [2];  [3];  [1];  [4];  [5];  [2];  [1];  [1];  [2];  [2];  [1];  [1];  [2];  [4]
  1. Fermi National Accelerator Lab. (FNAL), Batavia, IL (United States)
  2. European Organization for Nuclear Research (CERN), Geneva (Switzerland)
  3. Intel Corp., Santa Clara, CA (United States)
  4. Princeton Univ., Princeton, NJ (United States)
  5. Flatiron Institute of the Sions Foundation, New York, NY (United States)
Publication Date:
Research Org.:
Fermi National Accelerator Lab. (FNAL), Batavia, IL (United States)
Sponsoring Org.:
USDOE Office of Science (SC), High Energy Physics (HEP) (SC-25)
OSTI Identifier:
1414399
Report Number(s):
arXiv:1711.00375; FERMILAB-CONF-17-465-CD
Journal ID: ISSN 1742-6588; 1633859
Grant/Contract Number:  
AC02-07CH11359
Resource Type:
Accepted Manuscript
Journal Name:
Journal of Physics. Conference Series
Additional Journal Information:
Journal Volume: 1085; Journal Issue: 4; Conference: 18th International Workshop on Advanced Computing and Analysis Techniques in Physics Research, Seattle, WA (United States), 21-25 Aug 2017; Journal ID: ISSN 1742-6588
Publisher:
IOP Publishing
Country of Publication:
United States
Language:
English
Subject:
72 PHYSICS OF ELEMENTARY PARTICLES AND FIELDS

Citation Formats

Gutsche, Oliver, Canali, Luca, Cremer, Illia, Cremonesi, Matteo, Elmer, Peter, Fisk, Ian, Girone, Maria, Jayatilaka, Bo, Kowalkowski, Jim, Khristenko, Viktor, Motesnitsalis, Evangelos, Pivarski, Jim, Sehrish, Saba, Surdy, Kacper, and Svyatkovskiy, Alexey. CMS Analysis and Data Reduction with Apache Spark. United States: N. p., 2018. Web. doi:10.1088/1742-6596/1085/4/042030.
Gutsche, Oliver, Canali, Luca, Cremer, Illia, Cremonesi, Matteo, Elmer, Peter, Fisk, Ian, Girone, Maria, Jayatilaka, Bo, Kowalkowski, Jim, Khristenko, Viktor, Motesnitsalis, Evangelos, Pivarski, Jim, Sehrish, Saba, Surdy, Kacper, & Svyatkovskiy, Alexey. CMS Analysis and Data Reduction with Apache Spark. United States. doi:10.1088/1742-6596/1085/4/042030.
Gutsche, Oliver, Canali, Luca, Cremer, Illia, Cremonesi, Matteo, Elmer, Peter, Fisk, Ian, Girone, Maria, Jayatilaka, Bo, Kowalkowski, Jim, Khristenko, Viktor, Motesnitsalis, Evangelos, Pivarski, Jim, Sehrish, Saba, Surdy, Kacper, and Svyatkovskiy, Alexey. Thu . "CMS Analysis and Data Reduction with Apache Spark". United States. doi:10.1088/1742-6596/1085/4/042030. https://www.osti.gov/servlets/purl/1414399.
@article{osti_1414399,
title = {CMS Analysis and Data Reduction with Apache Spark},
author = {Gutsche, Oliver and Canali, Luca and Cremer, Illia and Cremonesi, Matteo and Elmer, Peter and Fisk, Ian and Girone, Maria and Jayatilaka, Bo and Kowalkowski, Jim and Khristenko, Viktor and Motesnitsalis, Evangelos and Pivarski, Jim and Sehrish, Saba and Surdy, Kacper and Svyatkovskiy, Alexey},
abstractNote = {Experimental Particle Physics has been at the forefront of analyzing the world's largest datasets for decades. The HEP community was among the first to develop suitable software and computing tools for this task. In recent times, new toolkits and systems for distributed data processing, collectively called "Big Data" technologies have emerged from industry and open source projects to support the analysis of Petabyte and Exabyte datasets in industry. While the principles of data analysis in HEP have not changed (filtering and transforming experiment-specific data formats), these new technologies use different approaches and tools, promising a fresh look at analysis of very large datasets that could potentially reduce the time-to-physics with increased interactivity. Moreover these new tools are typically actively developed by large communities, often profiting of industry resources, and under open source licensing. These factors result in a boost for adoption and maturity of the tools and for the communities supporting them, at the same time helping in reducing the cost of ownership for the end-users. In this talk, we are presenting studies of using Apache Spark for end user data analysis. We are studying the HEP analysis workflow separated into two thrusts: the reduction of centrally produced experiment datasets and the end-analysis up to the publication plot. Studying the first thrust, CMS is working together with CERN openlab and Intel on the CMS Big Data Reduction Facility. The goal is to reduce 1 PB of official CMS data to 1 TB of ntuple output for analysis. We are presenting the progress of this 2-year project with first results of scaling up Spark-based HEP analysis. Studying the second thrust, we are presenting studies on using Apache Spark for a CMS Dark Matter physics search, comparing Spark's feasibility, usability and performance to the ROOT-based analysis.},
doi = {10.1088/1742-6596/1085/4/042030},
journal = {Journal of Physics. Conference Series},
number = 4,
volume = 1085,
place = {United States},
year = {2018},
month = {10}
}

Journal Article:
Free Publicly Available Full Text
Publisher's Version of Record

Save / Share: