Calibration of Neural Networks using Splines

06/23/2020
by   Kartik Gupta, et al.
46

Calibrating neural networks is of utmost importance when employing them in safety-critical applications where the downstream decision making depends on the predicted probabilities. Measuring calibration error amounts to comparing two empirical distributions. In this work, we introduce a binning-free calibration measure inspired by the classical Kolmogorov-Smirnov (KS) statistical test in which the main idea is to compare the respective cumulative probability distributions. From this, by approximating the empirical cumulative distribution using a differentiable function via splines, we obtain a recalibration function, which maps the network outputs to actual (calibrated) class assignment probabilities. The spine-fitting is performed using a held-out calibration set and the obtained recalibration function is evaluated on an unseen test set. We tested our method against existing calibration approaches on various image classification datasets and our spline-based recalibration approach consistently outperforms existing methods on KS error as well as other commonly used calibration measures.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/22/2021

Localized Calibration: Metrics and Recalibration

Probabilistic classifiers output confidence scores along with their pred...
research
06/29/2020

Unsupervised Calibration under Covariate Shift

A probabilistic model is said to be calibrated if its predicted probabil...
research
07/12/2021

Calibrating Predictions to Decisions: A Novel Approach to Multi-Class Calibration

When facing uncertainty, decision-makers want predictions they can trust...
research
02/15/2022

Taking a Step Back with KCal: Multi-Class Kernel-Based Calibration for Deep Neural Networks

Deep neural network (DNN) classifiers are often overconfident, producing...
research
06/25/2023

TCE: A Test-Based Approach to Measuring Calibration Error

This paper proposes a new metric to measure the calibration error of pro...
research
10/28/2022

Stop Measuring Calibration When Humans Disagree

Calibration is a popular framework to evaluate whether a classifier know...
research
06/08/2023

Beyond Probability Partitions: Calibrating Neural Networks with Semantic Aware Grouping

Research has shown that deep networks tend to be overly optimistic about...

Please sign up or login with your details

Forgot password? Click here to reset