Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Beyond temperature scaling: Obtaining well-calibrated multiclass probabilities with Dirichlet calibration

About

Class probabilities predicted by most multiclass classifiers are uncalibrated, often tending towards over-confidence. With neural networks, calibration can be improved by temperature scaling, a method to learn a single corrective multiplicative factor for inputs to the last softmax layer. On non-neural models the existing methods apply binary calibration in a pairwise or one-vs-rest fashion. We propose a natively multiclass calibration method applicable to classifiers from any model class, derived from Dirichlet distributions and generalising the beta calibration method from binary classification. It is easily implemented with neural nets since it is equivalent to log-transforming the uncalibrated probabilities, followed by one linear layer and softmax. Experiments demonstrate improved probabilistic predictions according to multiple measures (confidence-ECE, classwise-ECE, log-loss, Brier score) across a wide range of datasets and classifiers. Parameters of the learned Dirichlet calibration map provide insights to the biases in the uncalibrated model.

Meelis Kull, Miquel Perello-Nieto, Markus K\"angsepp, Telmo Silva Filho, Hao Song, Peter Flach• 2019

Related benchmarks

TaskDatasetResultRank
Confidence calibrationDermatology
Confidence Calibration Error0.024
66
Model CalibrationCIFAR10 (test)--
61
Multi-class CalibrationCIFAR-100 logits (test)
LogLoss Absolute Improvement0.166
60
Confidence calibrationCAR
Calibration Error1.1
44
Confidence calibrationGlass
Calibration Error0.102
44
Confidence calibrationvehicle
Calibration Error0.063
44
Confidence calibrationCora
ECE0.0364
36
Confidence calibrationCiteseer
ECE4.87
36
Confidence calibrationPubmed
ECE0.0318
36
ClassificationGlass
Accuracy69.8
32
Showing 10 of 28 rows

Other info

Follow for update