Local Linearity and Double Descent in Catastrophic Overfitting

11/21/2021
by   Varun Sivashankar, et al.
0

Catastrophic overfitting is a phenomenon observed during Adversarial Training (AT) with the Fast Gradient Sign Method (FGSM) where the test robustness steeply declines over just one epoch in the training stage. Prior work has attributed this loss in robustness to a sharp decrease in local linearity of the neural network with respect to the input space, and has demonstrated that introducing a local linearity measure as a regularization term prevents catastrophic overfitting. Using a simple neural network architecture, we experimentally demonstrate that maintaining high local linearity might be sufficient to prevent catastrophic overfitting but is not necessary. Further, inspired by Parseval networks, we introduce a regularization term to AT with FGSM to make the weight matrices of the network orthogonal and study the connection between orthogonality of the network weights and local linearity. Lastly, we identify the double descent phenomenon during the adversarial training process.

READ FULL TEXT

page 1

page 2

research
10/05/2020

Understanding Catastrophic Overfitting in Single-step Adversarial Training

Adversarial examples are perturbed inputs that are designed to deceive m...
research
08/22/2023

Revisiting and Exploring Efficient Fast Adversarial Training via LAW: Lipschitz Regularization and Auto Weight Averaging

Fast Adversarial Training (FAT) not only improves the model robustness b...
research
07/06/2020

Understanding and Improving Fast Adversarial Training

A recent line of work focused on making adversarial training computation...
research
04/01/2023

Improving Fast Adversarial Training with Prior-Guided Knowledge

Fast adversarial training (FAT) is an efficient method to improve robust...
research
06/21/2023

Adversarial Training with Generated Data in High-Dimensional Regression: An Asymptotic Study

In recent years, studies such as <cit.> have demonstrated that incorpora...
research
10/07/2021

Double Descent in Adversarial Training: An Implicit Label Noise Perspective

Here, we show that the robust overfitting shall be viewed as the early p...

Please sign up or login with your details

Forgot password? Click here to reset