An Ensemble Approach Towards Adversarial Robustness

06/10/2021
by   Haifeng Qian, et al.
0

It is a known phenomenon that adversarial robustness comes at a cost to natural accuracy. To improve this trade-off, this paper proposes an ensemble approach that divides a complex robust-classification task into simpler subtasks. Specifically, fractal divide derives multiple training sets from the training data, and fractal aggregation combines inference outputs from multiple classifiers that are trained on those sets. The resulting ensemble classifiers have a unique property that ensures robustness for an input if certain don't-care conditions are met. The new techniques are evaluated on MNIST and Fashion-MNIST, with no adversarial training. The MNIST classifier has 99 natural accuracy, 70 L2 distance of 2. The Fashion-MNIST classifier has 90 measured robustness and 28.2 Both results are new state of the art, and we also present new state-of-the-art binary results on challenging label-pairs.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset