Denoising Multi-Source Weak Supervision for Neural Text Classification

10/09/2020
by   Wendi Ren, et al.
0

We study the problem of learning neural text classifiers without using any labeled data, but only easy-to-provide rules as multiple weak supervision sources. This problem is challenging because rule-induced weak labels are often noisy and incomplete. To address these two challenges, we design a label denoiser, which estimates the source reliability using a conditional soft attention mechanism and then reduces label noise by aggregating rule-annotated weak labels. The denoised pseudo labels then supervise a neural classifier to predicts soft labels for unmatched samples, which address the rule coverage issue. We evaluate our model on five benchmarks for sentiment, topic, and relation classifications. The results show that our model outperforms state-of-the-art weakly-supervised and semi-supervised methods consistently, and achieves comparable performance with fully-supervised methods even without any labeled data. Our code can be found at https://github.com/weakrules/Denoise-multi-weak-sources.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/07/2022

Label Propagation with Weak Supervision

Semi-supervised learning and weakly supervised learning are important pa...
research
04/13/2020

Learning from Rules Generalizing Labeled Exemplars

In many applications labeled data is not readily available, and needs to...
research
02/28/2022

Resolving label uncertainty with implicit posterior models

We propose a method for jointly inferring labels across a collection of ...
research
01/16/2021

Weakly-Supervised Hierarchical Models for Predicting Persuasive Strategies in Good-faith Textual Requests

Modeling persuasive language has the potential to better facilitate our ...
research
04/12/2021

Self-Training with Weak Supervision

State-of-the-art deep neural networks require large-scale labeled traini...
research
03/24/2022

Shoring Up the Foundations: Fusing Model Embeddings and Weak Supervision

Foundation models offer an exciting new paradigm for constructing models...
research
05/18/2022

Automatic Rule Induction for Efficient Semi-Supervised Learning

Semi-supervised learning has shown promise in allowing NLP models to gen...

Please sign up or login with your details

Forgot password? Click here to reset