Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

On Calibration of Modern Neural Networks

About

Confidence calibration -- the problem of predicting probability estimates representative of the true correctness likelihood -- is important for classification models in many applications. We discover that modern neural networks, unlike those from a decade ago, are poorly calibrated. Through extensive experiments, we observe that depth, width, weight decay, and Batch Normalization are important factors influencing calibration. We evaluate the performance of various post-processing calibration methods on state-of-the-art architectures with image and document classification datasets. Our analysis and experiments not only offer insights into neural network learning, but also provide a simple and straightforward recipe for practical settings: on most datasets, temperature scaling -- a single-parameter variant of Platt Scaling -- is surprisingly effective at calibrating predictions.

Chuan Guo, Geoff Pleiss, Yu Sun, Kilian Q. Weinberger• 2017

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-100
Top-1 Accuracy76.74
622
Image ClassificationFood-101
Accuracy86.6
494
Image ClassificationImageNet LT
Top-1 Accuracy37.9
251
Long-Tailed Image ClassificationImageNet-LT (test)--
220
Out-of-Distribution DetectioniNaturalist
FPR@9537.63
200
Image ClassificationImageNet-LT (test)--
159
Node ClassificationComputers--
143
Out-of-Distribution DetectionTextures
AUROC0.8539
141
Commonsense ReasoningARC Challenge
Accuracy64.9
132
Out-of-Distribution DetectionOpenImage-O
AUROC87.22
107
Showing 10 of 245 rows
...

Other info

Follow for update