Skip to main content
U.S. Department of Energy
Office of Scientific and Technical Information

Adaptive Neuron Apoptosis for Accelerating Deep Learning on Large Scale Systems

Conference ·

Machine Learning and Data Mining (MLDM) algorithms are becoming ubiquitous in {\em model learning} from the large volume of data generated using simulations, experiments and handheld devices. Deep Learning algorithms -- a class of MLDM algorithms -- are applied for automatic feature extraction, and learning non-linear models for unsupervised and supervised algorithms. Naturally, several libraries which support large scale Deep Learning -- such as TensorFlow and Caffe -- have become popular. In this paper, we present novel techniques to accelerate the convergence of Deep Learning algorithms by conducting low overhead removal of redundant neurons -- {\em apoptosis} of neurons -- which do not contribute to model learning, during the training phase itself. We provide in-depth theoretical underpinnings of our heuristics (bounding accuracy loss and handling apoptosis of several neuron types), and present the methods to conduct adaptive neuron apoptosis. We implement our proposed heuristics with the recently introduced TensorFlow and using its recently proposed extension with MPI. Our performance evaluation on two difference clusters -- one connected with Intel Haswell multi-core systems, and other with nVIDIA GPUs -- using InfiniBand, indicates the efficacy of the proposed heuristics and implementations. Specifically, we are able to improve the training time for several datasets by 2-3x, while reducing the number of parameters by 30x (4-5x on average) on datasets such as ImageNet classification. For the Higgs Boson dataset, our implementation improves the accuracy (measured by Area Under Curve (AUC)) for classification from 0.88/1 to 0.94/1, while reducing the number of parameters by 3x in comparison to existing literature, while achieving a 2.44x speedup in comparison to the default (no apoptosis) algorithm.

Research Organization:
Pacific Northwest National Laboratory (PNNL), Richland, WA (US)
Sponsoring Organization:
USDOE
DOE Contract Number:
AC05-76RL01830
OSTI ID:
1440683
Report Number(s):
PNNL-SA-120738; 453040300
Country of Publication:
United States
Language:
English

Similar Records

What does fault tolerant Deep Learning need from MPI?
Conference · Mon Sep 25 00:00:00 EDT 2017 · OSTI ID:1415701

Probing for Artifacts: Detecting Imagenet Model Evasions
Conference · Tue Jul 28 00:00:00 EDT 2020 · OSTI ID:1673321

Mitigating Catastrophic Forgetting in Deep Learning in a Streaming Setting Using Historical Summary
Conference · Mon Nov 01 00:00:00 EDT 2021 · OSTI ID:1885239

Related Subjects