Robust Neural Networks using Randomized Adversarial Training

03/25/2019
by   Alexandre Araujo, et al.
0

Since the discovery of adversarial examples in machine learning, researchers have designed several techniques to train neural networks that are robust against different types of attacks (most notably ℓ_∞ and ℓ_2 based attacks). However, it has been observed that the defense mechanisms designed to protect against one type of attack often offer poor performance against the other. In this paper, we introduce Randomized Adversarial Training (RAT), a technique that is efficient both against ℓ_2 and ℓ_∞ attacks. To obtain this result, we build upon adversarial training, a technique that is efficient against ℓ_∞ attacks, and demonstrate that adding random noise at training and inference time further improves performance against attacks. We then show that RAT is as efficient as adversarial training against ℓ_∞ attacks while being robust against strong ℓ_2 attacks. Our final comparative experiments demonstrate that RAT outperforms all state-of-the-art approaches against ℓ_2 and ℓ_∞ attacks.

READ FULL TEXT
research
12/01/2021

ℓ_∞-Robustness and Beyond: Unleashing Efficient Adversarial Training

Neural networks are vulnerable to adversarial attacks: adding well-craft...
research
05/18/2023

Re-thinking Data Availablity Attacks Against Deep Neural Networks

The unauthorized use of personal data for commercial purposes and the cl...
research
09/20/2019

Defending Against Physically Realizable Attacks on Image Classification

We study the problem of defending deep neural network approaches for ima...
research
06/05/2019

Enhancing Gradient-based Attacks with Symbolic Intervals

Recent breakthroughs in defenses against adversarial examples, like adve...
research
07/22/2021

Towards Explaining Adversarial Examples Phenomenon in Artificial Neural Networks

In this paper, we study the adversarial examples existence and adversari...
research
04/13/2022

A Novel Approach to Train Diverse Types of Language Models for Health Mention Classification of Tweets

Health mention classification deals with the disease detection in a give...
research
05/26/2019

Purifying Adversarial Perturbation with Adversarially Trained Auto-encoders

Machine learning models are vulnerable to adversarial examples. Iterativ...

Please sign up or login with your details

Forgot password? Click here to reset