skip to main content
OSTI.GOV title logo U.S. Department of Energy
Office of Scientific and Technical Information

Title: Graph interpolating activation improves both natural and robust accuracies in data-efficient deep learning

Journal Article · · European Journal of Applied Mathematics

Improving the accuracy and robustness of deep neural nets (DNNs) and adapting them to small training data are primary tasks in deep learning (DL) research. In this paper, we replace the output activation function of DNNs, typically the data-agnostic softmax function, with a graph Laplacian-based high-dimensional interpolating function which, in the continuum limit, converges to the solution of a Laplace–Beltrami equation on a high-dimensional manifold. Furthermore, we propose end-to-end training and testing algorithms for this new architecture. The proposed DNN with graph interpolating activation integrates the advantages of both deep learning and manifold learning. Compared to the conventional DNNs with the softmax function as output activation, the new framework demonstrates the following major advantages: First, it is better applicable to data-efficient learning in which we train high capacity DNNs without using a large number of training data. Second, it remarkably improves both natural accuracy on the clean images and robust accuracy on the adversarial images crafted by both white-box and black-box adversarial attacks. Third, it is a natural choice for semi-supervised learning. This paper is a significant extension of our earlier work published in NeurIPS, 2018. For reproducibility, the code is available athttps://github.com/BaoWangMath/DNN-DataDependentActivation.

Research Organization:
Purdue Univ., West Lafayette, IN (United States)
Sponsoring Organization:
USDOE Office of Science (SC)
DOE Contract Number:
SC0021142
OSTI ID:
1853729
Journal Information:
European Journal of Applied Mathematics, Vol. 32, Issue 3; ISSN 0956-7925
Publisher:
Cambridge University Press
Country of Publication:
United States
Language:
English

References (14)

Densely Connected Convolutional Networks conference July 2017
Identity Mappings in Deep Residual Networks book January 2016
Improving the Robustness of Deep Neural Networks via Stability Training conference June 2016
Distillation as a Defense to Adversarial Perturbations Against Deep Neural Networks conference May 2016
Scalable Nearest Neighbor Algorithms for High Dimensional Data journal November 2014
Wide Residual Networks conference January 2016
Universal Adversarial Perturbations conference July 2017
The Limitations of Deep Learning in Adversarial Settings conference March 2016
Deep Residual Learning for Image Recognition conference June 2016
Deep learning journal May 2015
A Fast Learning Algorithm for Deep Belief Nets journal July 2006
Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification conference December 2015
Certified Robustness to Adversarial Examples with Differential Privacy conference May 2019
Towards Evaluating the Robustness of Neural Networks conference May 2017

Similar Records

Improving Deep Neural Networks’ Training for Image Classification With Nonlinear Conjugate Gradient-Style Adaptive Momentum
Journal Article · Fri Mar 24 00:00:00 EDT 2023 · IEEE Transactions on Neural Networks and Learning Systems · OSTI ID:1853729

Towards automating structural discovery in scanning transmission electron microscopy *
Journal Article · Fri Feb 11 00:00:00 EST 2022 · Machine Learning: Science and Technology · OSTI ID:1853729

ASK: Adversarial Soft k-Nearest Neighbor Attack and Defense
Journal Article · Mon Sep 26 00:00:00 EDT 2022 · IEEE Access · OSTI ID:1853729

Related Subjects