Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Variational inference for Monte Carlo objectives

About

Recent progress in deep latent variable models has largely been driven by the development of flexible and scalable variational inference methods. Variational training of this type involves maximizing a lower bound on the log-likelihood, using samples from the variational posterior to compute the required gradients. Recently, Burda et al. (2016) have derived a tighter lower bound using a multi-sample importance sampling estimate of the likelihood and showed that optimizing it yields models that use more of their capacity and achieve higher likelihoods. This development showed the importance of such multi-sample objectives and explained the success of several related approaches. We extend the multi-sample approach to discrete latent variables and analyze the difficulty encountered when estimating the gradients involved. We then develop the first unbiased gradient estimator designed for importance-sampled objectives and evaluate it at training generative and structured output prediction models. The resulting estimator, which is based on low-variance per-sample learning signals, is both simpler and more effective than the NVIL estimator proposed for the single-sample variational objective, and is competitive with the currently used biased estimators.

Andriy Mnih, Danilo J. Rezende• 2016

Related benchmarks

TaskDatasetResultRank
Variational InferenceState-Space model (train)
Convergence Iterations298
18
Variational InferenceCOV
ELBO (nats)-8.09e+3
9
Variational InferenceDS2
ELBO (nats)-2.66e+4
9
Variational InferenceDS4
ELBO (nats)-1.35e+4
9
Variational InferenceDS11
ELBO (nats)-6.29e+3
9
Variational InferenceDS5
ELBO (nats)-8.28e+3
9
Variational InferenceDS9
ELBO (nats)-4.09e+3
9
Variational InferenceDS1
ELBO (nats)-7.16e+3
9
Variational InferenceDS3
ELBO (nats)-3.38e+4
9
Variational InferenceDS6
ELBO (nats)-6.76e+3
9
Showing 10 of 14 rows

Other info

Follow for update