Towards Interpretable Summary Evaluation via Allocation of Contextual Embeddings to Reference Text Topics

10/25/2022
by   Ben Schaper, et al.
0

Despite extensive recent advances in summary generation models, evaluation of auto-generated summaries still widely relies on single-score systems insufficient for transparent assessment and in-depth qualitative analysis. Towards bridging this gap, we propose the multifaceted interpretable summary evaluation method (MISEM), which is based on allocation of a summary's contextual token embeddings to semantic topics identified in the reference text. We further contribute an interpretability toolbox for automated summary evaluation and interactive visual analysis of summary scoring, topic identification, and token-topic allocation. MISEM achieves a promising .404 Pearson correlation with human judgment on the TAC'08 dataset.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/12/2021

Estimation of Summary-to-Text Inconsistency by Mismatched Embeddings

We propose a new reference-free summary quality evaluation measure, with...
research
08/28/2022

Podcast Summary Assessment: A Resource for Evaluating Summary Assessment Methods

Automatic summary assessment is useful for both machine-generated and hu...
research
11/22/2022

HaRiM^+: Evaluating Summary Quality with Hallucination Risk

One of the challenges of developing a summarization model arises from th...
research
07/11/2022

SummScore: A Comprehensive Evaluation Metric for Summary Quality Based on Cross-Encoder

Text summarization models are often trained to produce summaries that me...
research
02/18/2020

Learning by Semantic Similarity Makes Abstractive Summarization Better

One of the obstacles of abstractive summarization is the presence of var...
research
05/22/2023

Evaluating Factual Consistency of Texts with Semantic Role Labeling

Automated evaluation of text generation systems has recently seen increa...
research
10/07/2021

GeSERA: General-domain Summary Evaluation by Relevance Analysis

We present GeSERA, an open-source improved version of SERA for evaluatin...

Please sign up or login with your details

Forgot password? Click here to reset