Robust Positive-Unlabeled Learning via Noise Negative Sample Self-correction

08/01/2023
by   Zhangchi Zhu, et al.
1

Learning from positive and unlabeled data is known as positive-unlabeled (PU) learning in literature and has attracted much attention in recent years. One common approach in PU learning is to sample a set of pseudo-negatives from the unlabeled data using ad-hoc thresholds so that conventional supervised methods can be applied with both positive and negative samples. Owing to the label uncertainty among the unlabeled data, errors of misclassifying unlabeled positive samples as negative samples inevitably appear and may even accumulate during the training processes. Those errors often lead to performance degradation and model instability. To mitigate the impact of label uncertainty and improve the robustness of learning with positive and unlabeled data, we propose a new robust PU learning method with a training strategy motivated by the nature of human learning: easy cases should be learned first. Similar intuition has been utilized in curriculum learning to only use easier cases in the early stage of training before introducing more complex cases. Specifically, we utilize a novel “hardness” measure to distinguish unlabeled samples with a high chance of being negative from unlabeled samples with large label noise. An iterative training strategy is then implemented to fine-tune the selection of negative samples during the training process in an iterative manner to include more “easy” samples in the early stage of training. Extensive experimental validations over a wide range of learning tasks show that this approach can effectively improve the accuracy and stability of learning with positive and unlabeled data. Our code is available at https://github.com/woriazzc/Robust-PU

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/06/2022

Dist-PU: Positive-Unlabeled Learning from a Label Distribution Perspective

Positive-Unlabeled (PU) learning tries to learn binary classifiers from ...
research
01/29/2019

Revisiting Sample Selection Approach to Positive-Unlabeled Learning: Turning Unlabeled Data into Positive rather than Negative

In the early history of positive-unlabeled (PU) learning, the sample sel...
research
11/30/2022

Split-PU: Hardness-aware Training Strategy for Positive-Unlabeled Learning

Positive-Unlabeled (PU) learning aims to learn a model with rare positiv...
research
11/28/2016

Exploiting Unlabeled Data for Neural Grammatical Error Detection

Identifying and correcting grammatical errors in the text written by non...
research
02/13/2014

A Robust Ensemble Approach to Learn From Positive and Unlabeled Data Using SVM Base Models

We present a novel approach to learn binary classifiers when only positi...
research
09/14/2022

Improved proteasomal cleavage prediction with positive-unlabeled learning

Accurate in silico modeling of the antigen processing pathway is crucial...
research
11/21/2022

Uncertainty Reduction for 3D Point Cloud Self-Supervised Traversability Estimation

Traversability estimation in off-road environments requires a robust per...

Please sign up or login with your details

Forgot password? Click here to reset