S3: Supervised Self-supervised Learning under Label Noise

11/22/2021
by   Chen Feng, et al.
0

Despite the large progress in supervised learning with Neural Networks, there are significant challenges in obtaining high-quality, large-scale and accurately labeled datasets. In this context, in this paper we address the problem of classification in the presence of label noise and more specifically, both close-set and open-set label noise, that is when the true label of a sample may, or may not belong to the set of the given labels. In the heart of our method is a sample selection mechanism that relies on the consistency between the annotated label of a sample and the distribution of the labels in its neighborhood in the feature space; a relabeling mechanism that relies on the confidence of the classifier across subsequent iterations; and a training strategy that trains the encoder both with a self-consistency loss and the classifier-encoder with the cross-entropy loss on the selected samples alone. Without bells and whistles, such as co-training so as to reduce the self-confirmation bias, and with robustness with respect to settings of its few hyper-parameters, our method significantly surpasses previous methods on both CIFAR10/CIFAR100 with artificial noise and real-world noisy datasets such as WebVision and ANIMAL-10N.

READ FULL TEXT

page 2

page 3

research
11/16/2020

Decoupling Representation and Classifier for Noisy Label Learning

Since convolutional neural networks (ConvNets) can easily memorize noisy...
research
05/24/2023

SELFOOD: Self-Supervised Out-Of-Distribution Detection via Learning to Rank

Deep neural classifiers trained with cross-entropy loss (CE loss) often ...
research
10/04/2021

Consistency Regularization Can Improve Robustness to Label Noise

Consistency regularization is a commonly-used technique for semi-supervi...
research
03/28/2022

UNICON: Combating Label Noise Through Uniform Selection and Contrastive Learning

Supervised deep learning methods require a large repository of annotated...
research
03/16/2020

Neighborhood-based Pooling for Population-level Label Distribution Learning

Supervised machine learning often requires human-annotated data. While a...
research
04/20/2022

Quantity vs Quality: Investigating the Trade-Off between Sample Size and Label Reliability

In this paper, we study learning in probabilistic domains where the lear...
research
04/26/2020

Physics-constrained indirect supervised learning

This study proposes a supervised learning method that does not rely on l...

Please sign up or login with your details

Forgot password? Click here to reset