Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Calibration of Neural Networks using Splines

About

Calibrating neural networks is of utmost importance when employing them in safety-critical applications where the downstream decision making depends on the predicted probabilities. Measuring calibration error amounts to comparing two empirical distributions. In this work, we introduce a binning-free calibration measure inspired by the classical Kolmogorov-Smirnov (KS) statistical test in which the main idea is to compare the respective cumulative probability distributions. From this, by approximating the empirical cumulative distribution using a differentiable function via splines, we obtain a recalibration function, which maps the network outputs to actual (calibrated) class assignment probabilities. The spine-fitting is performed using a held-out calibration set and the obtained recalibration function is evaluated on an unseen test set. We tested our method against existing calibration approaches on various image classification datasets and our spline-based recalibration approach consistently outperforms existing methods on KS error as well as other commonly used calibration measures. Our Code is available at https://github.com/kartikgupta-at-anu/spline-calibration.

Kartik Gupta, Amir Rahimi, Thalaiyasingam Ajanthan, Thomas Mensink, Cristian Sminchisescu, Richard Hartley• 2020

Related benchmarks

TaskDatasetResultRank
Confidence calibrationCiteseer
ECE6.07
36
Confidence calibrationCora
ECE4.71
36
Confidence calibrationPubmed
ECE1.69
36
Confidence calibrationCoraFull
ECE2.68
28
GNN CalibrationComputers
ECE0.0156
12
GNN CalibrationCS
ECE1.08
12
GNN CalibrationPhysics
ECE0.45
12
GNN CalibrationPhoto
ECE1.59
12
Showing 8 of 8 rows

Other info

Follow for update