Quantifying the Task-Specific Information in Text-Based Classifications

10/17/2021
by   Zining Zhu, et al.
0

Recently, neural natural language models have attained state-of-the-art performance on a wide variety of tasks, but the high performance can result from superficial, surface-level cues (Bender and Koller, 2020; Niven and Kao, 2020). These surface cues, as the “shortcuts” inherent in the datasets, do not contribute to the *task-specific information* (TSI) of the classification tasks. While it is essential to look at the model performance, it is also important to understand the datasets. In this paper, we consider this question: Apart from the information introduced by the shortcut features, how much task-specific information is required to classify a dataset? We formulate this quantity in an information-theoretic framework. While this quantity is hard to compute, we approximate it with a fast and stable method. TSI quantifies the amount of linguistic knowledge modulo a set of predefined shortcuts – that contributes to classifying a sample from each dataset. This framework allows us to compare across datasets, saying that, apart from a set of “shortcut features”, classifying each sample in the Multi-NLI task involves around 0.4 nats more TSI than in the Quora Question Pair.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/19/2022

Are Prompt-based Models Clueless?

Finetuning large pre-trained language models with a task-specific head h...
research
04/06/2022

Improving Multi-task Generalization Ability for Neural Text Matching via Prompt Learning

Text matching is a fundamental technique in both information retrieval a...
research
05/09/2022

Attribution-based Task-specific Pruning for Multi-task Language Models

Multi-task language models show outstanding performance for various natu...
research
09/15/2023

Projected Task-Specific Layers for Multi-Task Reinforcement Learning

Multi-task reinforcement learning could enable robots to scale across a ...
research
03/07/2023

Exploring the Feasibility of ChatGPT for Event Extraction

Event extraction is a fundamental task in natural language processing th...
research
03/16/2022

TegTok: Augmenting Text Generation via Task-specific and Open-world Knowledge

Generating natural and informative texts has been a long-standing proble...
research
06/24/2020

Improving task-specific representation via 1M unlabelled images without any extra knowledge

We present a case-study to improve the task-specific representation by l...

Please sign up or login with your details

Forgot password? Click here to reset