DOE PAGES title logo U.S. Department of Energy
Office of Scientific and Technical Information

Title: Distilling particle knowledge for fast reconstruction at high-energy physics experiments

Journal Article · · Machine Learning: Science and Technology

Knowledge distillation is a form of model compression that allows artificial neural networks of different sizes to learn from one another. Its main application is the compactification of large deep neural networks to free up computational resources, in particular on edge devices. In this article, we consider proton-proton collisions at the High-Luminosity Large Hadron Collider (HL-LHC) and demonstrate a successful knowledge transfer from an event-level graph neural network (GNN) to a particle-level small deep neural network (DNN). Our algorithm, DistillNet, is a DNN that is trained to learn about the provenance of particles, as provided by the soft labels that are the GNN outputs, to predict whether or not a particle originates from the primary interaction vertex. The results indicate that for this problem, which is one of the main challenges at the HL-LHC, there is minimal loss during the transfer of knowledge to the small student network, while improving significantly the computational resource needs compared to the teacher. This is demonstrated for the distilled student network on a CPU, as well as for a quantized and pruned student network deployed on a field programmable gate array. Our study proves that knowledge transfer between networks of different complexity can be used for fast artificial intelligence (AI) in high-energy physics that improves the expressiveness of observables over non-AI-based reconstruction algorithms. Such an approach can become essential at the HL-LHC experiments, e.g. to comply with the resource budget of their trigger stages.

Research Organization:
Lawrence Berkeley National Laboratory (LBNL), Berkeley, CA (United States). National Energy Research Scientific Computing Center (NERSC)
Sponsoring Organization:
USDOE Office of Science (SC), High Energy Physics (HEP); USDOE Office of Science (SC), Basic Energy Sciences (BES). Scientific User Facilities (SUF); Swiss National Science Foundation (SNSF)
Grant/Contract Number:
AC02-05CH11231
OSTI ID:
2447590
Journal Information:
Machine Learning: Science and Technology, Journal Name: Machine Learning: Science and Technology Journal Issue: 2 Vol. 5; ISSN 2632-2153
Publisher:
IOP PublishingCopyright Statement
Country of Publication:
United States
Language:
English

References (18)

DELPHES 3: a modular framework for fast simulation of a generic collider experiment journal February 2014
Interleaved parton showers and tuning prospects journal March 2011
An efficient Lorentz equivariant graph neural network for jet tagging journal July 2022
Pileup per particle identification journal October 2014
An introduction to PYTHIA 8.2 journal June 2015
Automatic heterogeneous quantization of deep neural networks for low-latency inference on the edge for particle detectors journal June 2021
Particle Flow reconstruction in the CMS Level-1 Trigger For The HL-LHC journal January 2019
The anti- k t jet clustering algorithm journal April 2008
Fast inference of deep neural networks in FPGAs for particle physics journal July 2018
The ATLAS Experiment at the CERN Large Hadron Collider journal August 2008
The CMS experiment at the CERN LHC journal August 2008
Point cloud transformers applied to collider physics journal July 2021
Pile-up mitigation using attention journal June 2022
MLPF: efficient machine-learned particle-flow reconstruction using graph neural networks journal May 2021
Model compression and simplification pipelines for fast deep neural network inference in FPGAs in HEP journal November 2021
Semi-supervised graph neural networks for pileup noise removal journal January 2023
ABCNet: an attention-based method for particle tagging journal June 2020
On Information and Sufficiency journal March 1951