Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Stein Variational Gradient Descent: A General Purpose Bayesian Inference Algorithm

About

We propose a general purpose variational inference algorithm that forms a natural counterpart of gradient descent for optimization. Our method iteratively transports a set of particles to match the target distribution, by applying a form of functional gradient descent that minimizes the KL divergence. Empirical studies are performed on various real world models and datasets, on which our method is competitive with existing state-of-the-art methods. The derivation of our method is based on a new theoretical result that connects the derivative of KL divergence under smooth transforms with Stein's identity and a recently proposed kernelized Stein discrepancy, which is of independent interest.

Qiang Liu, Dilin Wang• 2016

Related benchmarks

TaskDatasetResultRank
Image ClassificationSVHN (test)--
362
Out-of-Distribution DetectionSVHN (test)
AUROC0.9355
48
Bayesian Neural NetworksUCI Boston (test)
RMSE2.774
10
Image ClassificationCIFAR-10 OOD: SVHN (test)
AUROC (H)0.825
10
Image ClassificationFashionMNIST OOD: MNIST (test)
AUROC (H)0.96
10
2D Synthetic Target SamplingGaussian 2D Synthetic
KSD0.013
8
2D Synthetic Target SamplingMOG2 2D Synthetic
KSD0.044
8
2D Synthetic Target SamplingROSENBROCK 2D Synthetic
KSD0.053
8
2D Synthetic Target SamplingDONUT 2D Synthetic
KSD0.057
8
2D Synthetic Target SamplingFUNNEL 2D Synthetic
KSD0.052
8
Showing 10 of 25 rows

Other info

Follow for update