Robustness Verification for Classifier Ensembles

05/12/2020
by   Dennis Gross, et al.
0

We give a formal verification procedure that decides whether a classifier ensemble is robust against arbitrary randomized attacks. Such attacks consist of a set of deterministic attacks and a distribution over this set. The robustness-checking problem consists of assessing, given a set of classifiers and a labelled data set, whether there exists a randomized attack that induces a certain expected loss against all classifiers. We show the NP-hardness of the problem and provide an upper bound on the number of attacks that is sufficient to form an optimal randomized attack. These results provide an effective way to reason about the robustness of a classifier ensemble. We provide SMT and MILP encodings to compute optimal randomized attacks or prove that there is no attack inducing a certain expected loss. In the latter case, the classifier ensemble is provably robust. Our prototype implementation verifies multiple neural-network ensembles trained for image-classification tasks. The experimental results using the MILP encoding are promising both in terms of scalability and the general applicability of our verification procedure.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/14/2023

Randomization for adversarial robustness: the Good, the Bad and the Ugly

Deep neural networks are known to be vulnerable to adversarial attacks: ...
research
05/26/2022

On Collective Robustness of Bagging Against Data Poisoning

Bootstrap aggregating (bagging) is an effective ensemble protocol, which...
research
12/09/2021

PARL: Enhancing Diversity of Ensemble Networks to Resist Adversarial Attacks via Pairwise Adversarially Robust Loss Function

The security of Deep Learning classifiers is a critical field of study b...
research
02/22/2021

On the robustness of randomized classifiers to adversarial examples

This paper investigates the theory of robustness against adversarial att...
research
03/27/2019

Scaling up the randomized gradient-free adversarial attack reveals overestimation of robustness using established attacks

Modern neural networks are highly non-robust against adversarial manipul...
research
04/25/2023

Certifying Ensembles: A General Certification Theory with S-Lipschitzness

Improving and guaranteeing the robustness of deep learning models has be...
research
02/21/2019

Learning requirements for stealth attacks

The learning data requirements are analyzed for the construction of stea...

Please sign up or login with your details

Forgot password? Click here to reset