Salvage Reusable Samples from Noisy Data for Robust Learning

08/06/2020
by   Zeren Sun, et al.
4

Due to the existence of label noise in web images and the high memorization capacity of deep neural networks, training deep fine-grained (FG) models directly through web images tends to have an inferior recognition ability. In the literature, to alleviate this issue, loss correction methods try to estimate the noise transition matrix, but the inevitable false correction would cause severe accumulated errors. Sample selection methods identify clean ("easy") samples based on the fact that small losses can alleviate the accumulated errors. However, "hard" and mislabeled examples that can both boost the robustness of FG models are also dropped. To this end, we propose a certainty-based reusable sample selection and correction approach, termed as CRSSC, for coping with label noise in training deep FG models with web images. Our key idea is to additionally identify and correct reusable samples, and then leverage them together with clean examples to update the networks. We demonstrate the superiority of the proposed approach from both theoretical and experimental perspectives.

READ FULL TEXT

page 1

page 2

page 4

page 5

page 6

page 7

page 8

page 9

research
08/06/2020

Data-driven Meta-set Based Fine-Grained Visual Classification

Constructing fine-grained image datasets typically requires domain-speci...
research
01/23/2021

Exploiting Web Images for Fine-Grained Visual Recognition by Eliminating Noisy Samples and Utilizing Hard Ones

Labeling objects at a subordinate level typically requires expert knowle...
research
08/23/2022

Learning from Noisy Labels with Coarse-to-Fine Sample Credibility Modeling

Training deep neural network (DNN) with noisy labels is practically chal...
research
07/21/2022

ProMix: Combating Label Noise via Maximizing Clean Sample Utility

The ability to train deep neural networks under label noise is appealing...
research
03/24/2021

Jo-SRC: A Contrastive Approach for Combating Noisy Labels

Due to the memorization effect in Deep Neural Networks (DNNs), training ...
research
02/14/2018

Using Trusted Data to Train Deep Networks on Labels Corrupted by Severe Noise

The growing importance of massive datasets with the advent of deep learn...
research
11/24/2018

Alternating Loss Correction for Preterm-Birth Prediction from EHR Data with Noisy Labels

In this paper we are interested in the prediction of preterm birth based...

Please sign up or login with your details

Forgot password? Click here to reset