Query Refinement Prompts for Closed-Book Long-Form Question Answering

10/31/2022
by   Reinald Kim Amplayo, et al.
0

Large language models (LLMs) have been shown to perform well in answering questions and in producing long-form texts, both in few-shot closed-book settings. While the former can be validated using well-known evaluation metrics, the latter is difficult to evaluate. We resolve the difficulties to evaluate long-form output by doing both tasks at once – to do question answering that requires long-form answers. Such questions tend to be multifaceted, i.e., they may have ambiguities and/or require information from multiple sources. To this end, we define query refinement prompts that encourage LLMs to explicitly express the multifacetedness in questions and generate long-form answers covering multiple facets of the question. Our experiments on two long-form question answering datasets, ASQA and AQuAMuSe, show that using our prompts allows us to outperform fully finetuned models in the closed book setting, as well as achieve results comparable to retrieve-then-generate open-book models.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/22/2021

Zero-Shot Open-Book Question Answering

Open book question answering is a subset of question answering tasks whe...
research
10/04/2021

Perhaps PTLMs Should Go to School – A Task to Assess Open Book and Closed Book QA

Our goal is to deliver a new task and leaderboard to stimulate research ...
research
05/21/2023

Model Analysis Evaluation for Ambiguous Question Answering

Ambiguous questions are a challenge for Question Answering models, as th...
research
05/29/2023

A Critical Evaluation of Evaluations for Long-form Question Answering

Long-form question answering (LFQA) enables answering a wide range of qu...
research
05/25/2023

The Dangers of trusting Stochastic Parrots: Faithfulness and Trust in Open-domain Conversational Question Answering

Large language models are known to produce output which sounds fluent an...
research
08/06/2020

Question and Answer Test-Train Overlap in Open-Domain Question Answering Datasets

Ideally Open-Domain Question Answering models should exhibit a number of...
research
12/26/2021

New Methods Metrics for LFQA tasks

Long-form question answering (LFQA) tasks require retrieving the documen...

Please sign up or login with your details

Forgot password? Click here to reset