Attribute-Guided Adversarial Training for Robustness to Natural Perturbations
Journal Article
·
· Proceedings of the AAAI Conference on Artificial Intelligence
- Arizona State University, Tempe, AZ (United States)
- Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States)
We report while existing work in robust deep learning has focused on small pixel-level norm-based perturbations, this may not account for perturbations encountered in several real world settings. In many such cases although test data might not be available, broad specifications about the types of perturbations (such as an unknown degree of rotation) may be known. We consider a setup where robustness is expected over an unseen test domain that is not i.i.d. but deviates from the training domain. While this deviation may not be exactly known, its broad characterization is specified a priori, in terms of attributes. We propose an adversarial training approach which learns to generate new samples so as to maximize exposure of the classifier to the attributes-space, without having access to the data from the test domain. Our adversarial training solves a min-max optimization problem, with the inner maximization generating adversarial perturbations, and the outer minimization finding model parameters by optimizing the loss on adversarial perturbations generated from the inner maximization. We demonstrate the applicability of our approach on three types of naturally occurring perturbations --- object-related shifts, geometric transformations, and common image corruptions. Our approach enables deep neural networks to be robust against a wide range of naturally occurring perturbations. We demonstrate the usefulness of the proposed approach by showing the robustness gains of deep neural networks trained using our adversarial training on MNIST, CIFAR-10, and a new variant of the CLEVR dataset.
- Research Organization:
- Lawrence Livermore National Laboratory (LLNL), Livermore, CA (United States)
- Sponsoring Organization:
- USDOE Laboratory Directed Research and Development (LDRD) Program; USDOE National Nuclear Security Administration (NNSA)
- Grant/Contract Number:
- AC52-07NA27344
- OSTI ID:
- 1888097
- Report Number(s):
- LLNL-JRNL-814425; 1023019
- Journal Information:
- Proceedings of the AAAI Conference on Artificial Intelligence, Journal Name: Proceedings of the AAAI Conference on Artificial Intelligence Journal Issue: 9 Vol. 35; ISSN 2159-5399
- Publisher:
- Association for the Advancement of Artificial IntelligenceCopyright Statement
- Country of Publication:
- United States
- Language:
- English
Similar Records
Exploring Robust Features for Improving Adversarial Robustness
Effects of Jacobian Matrix Regularization on the Detectability of Adversarial Samples
ASK: Adversarial Soft k-Nearest Neighbor Attack and Defense
Journal Article
·
Mon Apr 08 20:00:00 EDT 2024
· IEEE Transactions on Cybernetics
·
OSTI ID:2368822
Effects of Jacobian Matrix Regularization on the Detectability of Adversarial Samples
Technical Report
·
Tue Sep 01 00:00:00 EDT 2020
·
OSTI ID:1763568
ASK: Adversarial Soft k-Nearest Neighbor Attack and Defense
Journal Article
·
Sun Sep 25 20:00:00 EDT 2022
· IEEE Access
·
OSTI ID:1922471