Co-Mixup: Saliency Guided Joint Mixup with Supermodular Diversity

02/05/2021
by   Jang-Hyun Kim, et al.
6

While deep neural networks show great performance on fitting to the training distribution, improving the networks' generalization performance to the test distribution and robustness to the sensitivity to input perturbations still remain as a challenge. Although a number of mixup based augmentation strategies have been proposed to partially address them, it remains unclear as to how to best utilize the supervisory signal within each input data for mixup from the optimization perspective. We propose a new perspective on batch mixup and formulate the optimal construction of a batch of mixup data maximizing the data saliency measure of each individual mixup data and encouraging the supermodular diversity among the constructed mixup data. This leads to a novel discrete optimization problem minimizing the difference between submodular functions. We also propose an efficient modular approximation based iterative submodular minimization algorithm for efficient mixup computation per each minibatch suitable for minibatch based neural network training. Our experiments show the proposed method achieves the state of the art generalization, calibration, and weakly supervised localization results compared to other mixup methods. The source code is available at https://github.com/snu-mllab/Co-Mixup.

READ FULL TEXT

page 2

page 18

page 19

page 20

page 21

research
09/15/2020

Puzzle Mix: Exploiting Saliency and Local Statistics for Optimal Mixup

While deep neural networks achieve great performance on fitting the trai...
research
12/09/2022

Expeditious Saliency-guided Mix-up through Random Gradient Thresholding

Mix-up training approaches have proven to be effective in improving the ...
research
03/14/2020

Investigating Generalization in Neural Networks under Optimally Evolved Training Perturbations

In this paper, we study the generalization properties of neural networks...
research
11/28/2020

GradAug: A New Regularization Method for Deep Neural Networks

We propose a new regularization method to alleviate over-fitting in deep...
research
02/22/2022

Submodlib: A Submodular Optimization Library

Submodular functions are a special class of set functions which naturall...
research
05/29/2021

Greedy Bayesian Posterior Approximation with Deep Ensembles

Ensembles of independently trained neural networks are a state-of-the-ar...
research
07/13/2021

Fast Batch Nuclear-norm Maximization and Minimization for Robust Domain Adaptation

Due to the domain discrepancy in visual domain adaptation, the performan...

Please sign up or login with your details

Forgot password? Click here to reset