Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Accurate Uncertainties for Deep Learning Using Calibrated Regression

About

Methods for reasoning under uncertainty are a key building block of accurate and reliable machine learning systems. Bayesian methods provide a general framework to quantify uncertainty. However, because of model misspecification and the use of approximate inference, Bayesian uncertainty estimates are often inaccurate -- for example, a 90% credible interval may not contain the true outcome 90% of the time. Here, we propose a simple procedure for calibrating any regression algorithm; when applied to Bayesian and probabilistic models, it is guaranteed to produce calibrated uncertainty estimates given enough data. Our procedure is inspired by Platt scaling and extends previous work on classification. We evaluate this approach on Bayesian linear regression, feedforward, and recurrent neural networks, and find that it consistently outputs well-calibrated credible intervals while improving performance on time series forecasting and model-based reinforcement learning tasks.

Volodymyr Kuleshov, Nathan Fenner, Stefano Ermon• 2018

Related benchmarks

TaskDatasetResultRank
RegressionBoston UCI (test)--
26
RegressionConcrete UCI (test pool)
MACE0.044
14
Uncertainty CalibrationAuto-MPG MPG2
MACE0.059
6
Uncertainty CalibrationConcrete
MACE0.059
6
Uncertainty CalibrationMEPS Panel 20 2017 (test)
Length6.94e+3
6
Uncertainty CalibrationAuto-MPG MPG3
MACE0.1
6
Uncertainty CalibrationBoston
MACE11
6
Uncertainty CalibrationWine
MACE0.31
6
Uncertainty CalibrationMEPS Panel 19 2017 (test)
Length506
6
Uncertainty CalibrationMEPS Panel 21 2017 (test)
Length4.32e+3
6
Showing 10 of 10 rows

Other info

Follow for update