LLM Calibration and Automatic Hallucination Detection via Pareto Optimal Self-supervision

06/28/2023
by   Theodore Zhao, et al.
0

Large language models (LLMs) have demonstrated remarkable capabilities out of box for a wide range of applications, yet accuracy still remains a major growth area, especially in mission-critical domains such as biomedicine. An effective method to calibrate the confidence level on LLM responses is essential to automatically detect errors and facilitate human-in-the-loop verification. An important source of calibration signals stems from expert-stipulated programmatic supervision, which is often available at low cost but has its own limitations such as noise and coverage. In this paper, we introduce a Pareto optimal self-supervision framework that can leverage available programmatic supervision to systematically calibrate LLM responses by producing a risk score for every response, without any additional manual efforts. This is accomplished by learning a harmonizer model to align LLM output with other available supervision sources, which would assign higher risk scores to more uncertain LLM responses and facilitate error correction. Experiments on standard relation extraction tasks in biomedical and general domains demonstrate the promise of this approach, with our proposed risk scores highly correlated with the real error rate of LLMs. For the most uncertain test instances, dynamic prompting based on our proposed risk scores results in significant accuracy improvement for off-the-shelf LLMs, boosting GPT-3 results past state-of-the-art (SOTA) weak supervision and GPT-4 results past SOTA supervised results on challenging evaluation datasets.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/11/2021

Modular Self-Supervision for Document-Level Relation Extraction

Extracting relations across large text spans has been relatively underex...
research
05/11/2017

Learning with Noise: Enhance Distantly Supervised Relation Extraction with Dynamic Transition Matrix

Distant supervision significantly reduces human efforts in building trai...
research
07/31/2022

Improving Distantly Supervised Relation Extraction by Natural Language Inference

To reduce human annotations for relation extraction (RE) tasks, distantl...
research
07/12/2023

Distilling Large Language Models for Biomedical Knowledge Extraction: A Case Study on Adverse Drug Events

Large language models (LLMs), such as GPT-4, have demonstrated remarkabl...
research
04/28/2021

Multi-view Inference for Relation Extraction with Uncertain Knowledge

Knowledge graphs (KGs) are widely used to facilitate relation extraction...
research
08/06/2023

Empirical Optimal Risk to Quantify Model Trustworthiness for Failure Detection

Failure detection (FD) in AI systems is a crucial safeguard for the depl...

Please sign up or login with your details

Forgot password? Click here to reset