Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Variational Dropout and the Local Reparameterization Trick

About

We investigate a local reparameterizaton technique for greatly reducing the variance of stochastic gradients for variational Bayesian inference (SGVB) of a posterior over model parameters, while retaining parallelizability. This local reparameterization translates uncertainty about global parameters into local noise that is independent across datapoints in the minibatch. Such parameterizations can be trivially parallelized and have variance that is inversely proportional to the minibatch size, generally leading to much faster convergence. Additionally, we explore a connection with dropout: Gaussian dropout objectives correspond to SGVB with local reparameterization, a scale-invariant prior and proportionally fixed posterior variance. Our method allows inference of more flexibly parameterized posteriors; specifically, we propose variational dropout, a generalization of Gaussian dropout where the dropout rates are learned, often leading to better models. The method is demonstrated through several experiments.

Diederik P. Kingma, Tim Salimans, Max Welling• 2015

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-100--
622
Image ClassificationSVHN (test)
Accuracy90.76
362
Image ClassificationSVHN
Accuracy96.11
359
Image ClassificationSTL-10 (test)
Accuracy65.48
357
Image ClassificationCIFAR100 (test)
Accuracy43.1
112
RegressionUCI ENERGY (test)
Negative Log Likelihood1.3
42
RegressionUCI CONCRETE (test)
Neg Log Likelihood-3.07
37
RegressionUCI YACHT (test)
Negative Log Likelihood-1.42
33
RegressionUCI POWER (test)
Negative Log Likelihood-2.82
29
RegressionEnergy UCI (test)
RMSE0.64
27
Showing 10 of 28 rows

Other info

Follow for update