Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Practical Deep Learning with Bayesian Principles

About

Bayesian methods promise to fix many shortcomings of deep learning, but they are impractical and rarely match the performance of standard methods, let alone improve them. In this paper, we demonstrate practical training of deep networks with natural-gradient variational inference. By applying techniques such as batch normalisation, data augmentation, and distributed training, we achieve similar performance in about the same number of epochs as the Adam optimiser, even on large datasets such as ImageNet. Importantly, the benefits of Bayesian principles are preserved: predictive probabilities are well-calibrated, uncertainties on out-of-distribution data are improved, and continual-learning performance is boosted. This work enables practical deep learning while preserving benefits of Bayesian principles. A PyTorch implementation is available as a plug-and-play optimiser.

Kazuki Osawa, Siddharth Swaroop, Anirudh Jain, Runa Eschenhagen, Richard E. Turner, Rio Yokota, Mohammad Emtiyaz Khan• 2019

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-10
Accuracy84.27
507
Image ClassificationSVHN (test)--
362
Out-of-Distribution DetectionSVHN
AUROC87.6
62
Image ClassificationCIFAR10 (test)
Error Rate26.91
19
Image ClassificationMNIST vectorized (test)
NLL0.049
15
Showing 5 of 5 rows

Other info

Follow for update