Learning Imbalanced Datasets with Label-Distribution-Aware Margin Loss

06/18/2019
by   Kaidi Cao, et al.
1

Deep learning algorithms can fare poorly when the training dataset suffers from heavy class-imbalance but the testing criterion requires good generalization on less frequent classes. We design two novel methods to improve performance in such scenarios. First, we propose a theoretically-principled label-distribution-aware margin (LDAM) loss motivated by minimizing a margin-based generalization bound. This loss replaces the standard cross-entropy objective during training and can be applied with prior strategies for training with class-imbalance such as re-weighting or re-sampling. Second, we propose a simple, yet effective, training schedule that defers re-weighting until after the initial stage, allowing the model to learn an initial representation while avoiding some of the complications associated with re-weighting or re-sampling. We test our methods on several benchmark vision tasks including the real-world imbalanced dataset iNaturalist 2018. Our experiments show that either of these methods alone can already improve over existing techniques and their combination achieves even better performance gains.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/11/2022

Learning Imbalanced Datasets with Maximum Margin Loss

A learning algorithm referred to as Maximum Margin (MM) is proposed for ...
research
12/28/2022

Escaping Saddle Points for Effective Generalization on Class-Imbalanced Data

Real-world datasets exhibit imbalances of varying types and degrees. Sev...
research
10/10/2022

A Survey of Methods for Addressing Class Imbalance in Deep-Learning Based Natural Language Processing

Many natural language processing (NLP) tasks are naturally imbalanced, a...
research
11/25/2020

Supercharging Imbalanced Data Learning With Causal Representation Transfer

Dealing with severe class imbalance poses a major challenge for real-wor...
research
06/25/2022

On how to avoid exacerbating spurious correlations when models are overparameterized

Overparameterized models fail to generalize well in the presence of data...
research
06/23/2022

Prototype-Anchored Learning for Learning with Imperfect Annotations

The success of deep neural networks greatly relies on the availability o...
research
01/23/2019

Max-margin Class Imbalanced Learning with Gaussian Affinity

Real-world object classes appear in imbalanced ratios. This poses a sign...

Please sign up or login with your details

Forgot password? Click here to reset