Hidden Heterogeneity: When to Choose Similarity-Based Calibration

02/03/2022
by   Kiri L. Wagstaff, et al.
7

Trustworthy classifiers are essential to the adoption of machine learning predictions in many real-world settings. The predicted probability of possible outcomes can inform high-stakes decision making, particularly when assessing the expected value of alternative decisions or the risk of bad outcomes. These decisions require well calibrated probabilities, not just the correct prediction of the most likely class. Black-box classifier calibration methods can improve the reliability of a classifier's output without requiring retraining. However, these methods are unable to detect subpopulations where calibration could improve prediction accuracy. Such subpopulations are said to exhibit "hidden heterogeneity" (HH), because the original classifier did not detect them. The paper proposes a quantitative measure for HH. It also introduces two similarity-weighted calibration methods that can address HH by adapting locally to each test item: SWC weights the calibration set by similarity to the test item, and SWC-HH explicitly incorporates hidden heterogeneity to filter the calibration set. Experiments show that the improvements in calibration achieved by similarity-based calibration methods correlate with the amount of HH present and, given sufficient calibration data, generally exceed calibration achieved by global methods. HH can therefore serve as a useful diagnostic tool for identifying when local calibration methods are needed.

READ FULL TEXT

page 15

page 17

page 19

research
04/02/2019

Measuring Calibration in Deep Learning

The reliability of a machine learning model's confidence in its predicti...
research
05/23/2022

What is Your Metric Telling You? Evaluating Classifier Calibration under Context-Specific Definitions of Reliability

Classifier calibration has received recent attention from the machine le...
research
05/05/2023

Calibration Assessment and Boldness-Recalibration for Binary Events

Probability predictions are essential to inform decision making in medic...
research
07/12/2021

Calibrating Predictions to Decisions: A Novel Approach to Multi-Class Calibration

When facing uncertainty, decision-makers want predictions they can trust...
research
12/20/2021

Classifier Calibration: How to assess and improve predicted class probabilities: a survey

This paper provides both an introduction to and a detailed overview of t...
research
07/27/2022

Calibrate: Interactive Analysis of Probabilistic Model Output

Analyzing classification model performance is a crucial task for machine...
research
03/18/2022

Decision-Making under Miscalibration

ML-based predictions are used to inform consequential decisions about in...

Please sign up or login with your details

Forgot password? Click here to reset