Domain Aligned Prefix Averaging for Domain Generalization in Abstractive Summarization

05/26/2023
by   Pranav Ajit Nair, et al.
0

Domain generalization is hitherto an underexplored area applied in abstractive summarization. Moreover, most existing works on domain generalization have sophisticated training algorithms. In this paper, we propose a lightweight, weight averaging based, Domain Aligned Prefix Averaging approach to domain generalization for abstractive summarization. Given a number of source domains, our method first trains a prefix for each one of them. These source prefixes generate summaries for a small number of target domain documents. The similarity of the generated summaries to their corresponding documents is used for calculating weights required to average source prefixes. In DAPA, prefix tuning allows for lightweight finetuning, and weight averaging allows for the computationally efficient addition of new source domains. When evaluated on four diverse summarization domains, DAPA shows comparable or better performance against the baselines, demonstrating the effectiveness of its prefix averaging scheme.

READ FULL TEXT
research
02/17/2021

Domain Generalization Needs Stochastic Weight Averaging for Robustness on Domain Shifts

Domain generalization aims to learn a generalizable model to unseen targ...
research
07/21/2017

A Pilot Study of Domain Adaptation Effect for Neural Abstractive Summarization

We study the problem of domain adaptation for neural abstractive summari...
research
05/25/2018

Reinforced Extractive Summarization with Question-Focused Rewards

We investigate a new training paradigm for extractive summarization. Tra...
research
02/14/2023

AdapterSoup: Weight Averaging to Improve Generalization of Pretrained Language Models

Pretrained language models (PLMs) are trained on massive corpora, but of...
research
11/15/2020

Open4Business(O4B): An Open Access Dataset for Summarizing Business Documents

A major challenge in fine-tuning deep learning models for automatic summ...
research
11/14/2016

SummaRuNNer: A Recurrent Neural Network based Sequence Model for Extractive Summarization of Documents

We present SummaRuNNer, a Recurrent Neural Network (RNN) based sequence ...
research
02/28/2023

DART: Diversify-Aggregate-Repeat Training Improves Generalization of Neural Networks

Generalization of neural networks is crucial for deploying them safely i...

Please sign up or login with your details

Forgot password? Click here to reset