Reachable Sets of Classifiers Regression Models: (Non-)Robustness Analysis and Robust Training

07/28/2020
by   Anna-Kathrin Kopetzki, et al.
0

Neural networks achieve outstanding accuracy in classification and regression tasks. However, understanding their behavior still remains an open challenge that requires questions to be addressed on the robustness, explainability and reliability of predictions. We answer these questions by computing reachable sets of neural networks, i.e. sets of outputs resulting from continuous sets of inputs. We provide two efficient approaches that lead to over- and under-approximations of the reachable set. This principle is highly versatile, as we show. First, we analyze and enhance the robustness properties of both classifiers and regression models. This is in contrast to existing works, which only handle classification. Specifically, we verify (non-)robustness, propose a robust training procedure, and show that our approach outperforms adversarial attacks as well as state-of-the-art methods of verifying classifiers for non-norm bound perturbations. We also provide a technique of distinguishing between reliable and non-reliable predictions for unlabeled inputs, quantify the influence of each feature on a prediction, and compute a feature ranking.

READ FULL TEXT

page 17

page 19

research
06/08/2020

Adversarial Feature Desensitization

Deep neural networks can now perform many tasks that were once thought t...
research
04/22/2020

Adversarial examples and where to find them

Adversarial robustness of trained models has attracted considerable atte...
research
05/25/2023

Don't Retrain, Just Rewrite: Countering Adversarial Perturbations by Rewriting Text

Can language models transform inputs to protect text classifiers against...
research
07/24/2022

Can we achieve robustness from data alone?

Adversarial training and its variants have come to be the prevailing met...
research
10/31/2019

Certifiable Robustness to Graph Perturbations

Despite the exploding interest in graph neural networks there has been l...
research
03/24/2023

Feature Separation and Recalibration for Adversarial Robustness

Deep neural networks are susceptible to adversarial attacks due to the a...
research
05/03/2022

On the uncertainty principle of neural networks

Despite the successes in many fields, it is found that neural networks a...

Please sign up or login with your details

Forgot password? Click here to reset