Evaluating Paraphrastic Robustness in Textual Entailment Models

06/29/2023
by   Dhruv Verma, et al.
0

We present PaRTE, a collection of 1,126 pairs of Recognizing Textual Entailment (RTE) examples to evaluate whether models are robust to paraphrasing. We posit that if RTE models understand language, their predictions should be consistent across inputs that share the same meaning. We use the evaluation set to determine if RTE models' predictions change when examples are paraphrased. In our experiments, contemporary models change their predictions on 8-16% of paraphrased examples, indicating that there is still room for improvement.

READ FULL TEXT
research
06/02/2021

Figurative Language in Recognizing Textual Entailment

We introduce a collection of recognizing textual entailment (RTE) datase...
research
10/25/2018

Teaching Syntax by Adversarial Distraction

Existing entailment datasets mainly pose problems which can be answered ...
research
05/12/2018

AdvEntuRe: Adversarial Training for Textual Entailment with Knowledge-Guided Examples

We consider the problem of learning textual entailment models with limit...
research
06/10/2018

What Knowledge is Needed to Solve the RTE5 Textual Entailment Challenge?

This document gives a knowledge-oriented analysis of about 20 interestin...
research
10/06/2020

A Survey on Recognizing Textual Entailment as an NLP Evaluation

Recognizing Textual Entailment (RTE) was proposed as a unified evaluatio...
research
08/28/2018

Bridging Knowledge Gaps in Neural Entailment via Symbolic Models

Most textual entailment models focus on lexical gaps between the premise...
research
04/20/2018

Acquisition of Phrase Correspondences using Natural Deduction Proofs

How to identify, extract, and use phrasal knowledge is a crucial problem...

Please sign up or login with your details

Forgot password? Click here to reset