Noised Consistency Training for Text Summarization

05/28/2021
by   Junnan Liu, et al.
0

Neural abstractive summarization methods often require large quantities of labeled training data. However, labeling large amounts of summarization data is often prohibitive due to time, financial, and expertise constraints, which has limited the usefulness of summarization systems to practical applications. In this paper, we argue that this limitation can be overcome by a semi-supervised approach: consistency training which is to leverage large amounts of unlabeled data to improve the performance of supervised learning over a small corpus. The consistency regularization semi-supervised learning can regularize model predictions to be invariant to small noise applied to input articles. By adding noised unlabeled corpus to help regularize consistency training, this framework obtains comparative performance without using the full dataset. In particular, we have verified that leveraging large amounts of unlabeled data decently improves the performance of supervised learning over an insufficient labeled dataset.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/31/2020

Pseudo-Representation Labeling Semi-Supervised Learning

In recent years, semi-supervised learning (SSL) has shown tremendous suc...
research
04/08/2020

Empirical Perspectives on One-Shot Semi-supervised Learning

One of the greatest obstacles in the adoption of deep neural networks fo...
research
03/04/2020

Semixup: In- and Out-of-Manifold Regularization for Deep Semi-Supervised Knee Osteoarthritis Severity Grading from Plain Radiographs

Knee osteoarthritis (OA) is one of the highest disability factors in the...
research
02/01/2022

Semi-supervised 3D Object Detection via Temporal Graph Neural Networks

3D object detection plays an important role in autonomous driving and ot...
research
11/01/2022

The Perils of Learning From Unlabeled Data: Backdoor Attacks on Semi-supervised Learning

Semi-supervised machine learning (SSL) is gaining popularity as it reduc...
research
10/27/2020

To BERT or Not to BERT: Comparing Task-specific and Task-agnostic Semi-Supervised Approaches for Sequence Tagging

Leveraging large amounts of unlabeled data using Transformer-like archit...
research
04/21/2022

SelfD: Self-Learning Large-Scale Driving Policies From the Web

Effectively utilizing the vast amounts of ego-centric navigation data th...

Please sign up or login with your details

Forgot password? Click here to reset