Detecting Hallucinated Content in Conditional Neural Sequence Generation

11/05/2020
by   Chunting Zhou, et al.
2

Neural sequence models can generate highly fluent sentences but recent studies have also shown that they are also prone to hallucinate additional content not supported by the input, which can cause a lack of trust in the model. To better assess the faithfulness of the machine outputs, we propose a new task to predict whether each token in the output sequence is hallucinated conditioned on the source input, and collect new manually annotated evaluation sets for this task. We also introduce a novel method for learning to model hallucination detection, based on pretrained language models fine tuned on synthetic data that includes automatically inserted hallucinations. Experiments on machine translation and abstract text summarization demonstrate the effectiveness of our proposed approach – we obtain an average F1 of around 0.6 across all the benchmark datasets and achieve significant improvements in sentence-level hallucination scoring compared to baseline methods. We also release our annotated data and code for future research at https://github.com/violet-zct/fairseq-detect-hallucination.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/19/2023

The Inside Story: Towards Better Understanding of Machine Translation Neural Evaluation Metrics

Neural metrics for machine translation evaluation, such as COMET, exhibi...
research
11/08/2022

Conciseness: An Overlooked Language Task

We report on novel investigations into training models that make sentenc...
research
08/23/2023

Evaluation of Faithfulness Using the Longest Supported Subsequence

As increasingly sophisticated language models emerge, their trustworthin...
research
09/15/2023

Large Language Models for Failure Mode Classification: An Investigation

In this paper we present the first investigation into the effectiveness ...
research
05/29/2021

CoDesc: A Large Code-Description Parallel Dataset

Translation between natural language and source code can help software d...
research
08/29/2022

Reweighting Strategy based on Synthetic Data Identification for Sentence Similarity

Semantically meaningful sentence embeddings are important for numerous t...
research
09/30/2022

Out-of-Distribution Detection and Selective Generation for Conditional Language Models

Machine learning algorithms typically assume independent and identically...

Please sign up or login with your details

Forgot password? Click here to reset