Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Accelerated Information Gradient flow

About

We present a framework for Nesterov's accelerated gradient flows in probability space to design efficient mean-field Markov chain Monte Carlo (MCMC) algorithms for Bayesian inverse problems. Here four examples of information metrics are considered, including Fisher-Rao metric, Wasserstein-2 metric, Kalman-Wasserstein metric and Stein metric. For both Fisher-Rao and Wasserstein-2 metrics, we prove convergence properties of accelerated gradient flows. In implementations, we propose a sampling-efficient discrete-time algorithm for Wasserstein-2, Kalman-Wasserstein and Stein accelerated gradient flows with a restart technique. We also formulate a kernel bandwidth selection method, which learns the gradient of logarithm of density from Brownian-motion samples. Numerical experiments, including Bayesian logistic regression and Bayesian neural network, show the strength of the proposed methods compared with state-of-the-art algorithms.

Yifei Wang, Wuchen Li• 2019

Related benchmarks

TaskDatasetResultRank
Bayesian Neural NetworksUCI Boston (test)
RMSE3.077
10
Bayesian Neural Network RegressionWINE (test)
RMSE0.614
6
Bayesian Neural Network Regressionconcrete (test)
RMSE4.883
6
Bayesian Neural Network RegressionCombined (test)
RMSE4.077
6
Bayesian Neural Network Regressionkin8nm (test)
RMSE0.096
6
Showing 5 of 5 rows

Other info

Follow for update