Few-shot Text Classification with Dual Contrastive Consistency

09/29/2022
by   Liwen Sun, et al.
0

In this paper, we explore how to utilize pre-trained language model to perform few-shot text classification where only a few annotated examples are given for each class. Since using traditional cross-entropy loss to fine-tune language model under this scenario causes serious overfitting and leads to sub-optimal generalization of model, we adopt supervised contrastive learning on few labeled data and consistency-regularization on vast unlabeled data. Moreover, we propose a novel contrastive consistency to further boost model performance and refine sentence representation. After conducting extensive experiments on four datasets, we demonstrate that our model (FTCC) can outperform state-of-the-art methods and has better robustness.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/03/2020

Supervised Contrastive Learning for Pre-trained Language Model Fine-tuning

State-of-the-art natural language understanding classification models fo...
research
10/29/2022

Differentiable Data Augmentation for Contrastive Sentence Representation Learning

Fine-tuning a pre-trained language model via the contrastive learning fr...
research
10/23/2022

Discriminative Language Model as Semantic Consistency Scorer for Prompt-based Few-Shot Text Classification

This paper proposes a novel prompt-based finetuning method (called DLM-S...
research
12/21/2021

Supervised Graph Contrastive Pretraining for Text Classification

Contrastive pretraining techniques for text classification has been larg...
research
04/11/2021

Constructing Contrastive samples via Summarization for Text Classification with limited annotations

Contrastive Learning has emerged as a powerful representation learning m...
research
08/28/2023

Breaking the Bank with ChatGPT: Few-Shot Text Classification for Finance

We propose the use of conversational GPT models for easy and quick few-s...
research
10/04/2021

Revisiting Self-Training for Few-Shot Learning of Language Model

As unlabeled data carry rich task-relevant information, they are proven ...

Please sign up or login with your details

Forgot password? Click here to reset