Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

No MCMC for me: Amortized sampling for fast and stable training of energy-based models

About

Energy-Based Models (EBMs) present a flexible and appealing way to represent uncertainty. Despite recent advances, training EBMs on high-dimensional data remains a challenging problem as the state-of-the-art approaches are costly, unstable, and require considerable tuning and domain expertise to apply successfully. In this work, we present a simple method for training EBMs at scale which uses an entropy-regularized generator to amortize the MCMC sampling typically used in EBM training. We improve upon prior MCMC-based entropy regularization methods with a fast variational approximation. We demonstrate the effectiveness of our approach by using it to train tractable likelihood models. Next, we apply our estimator to the recently proposed Joint Energy Model (JEM), where we match the original performance with faster and stable training. This allows us to extend JEM models to semi-supervised classification on tabular data from a variety of continuous domains.

Will Grathwohl, Jacob Kelly, Milad Hashemi, Mohammad Norouzi, Kevin Swersky, David Duvenaud• 2020

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR10 (test)
Accuracy93.2
585
Image GenerationCIFAR-10 (test)
FID30.5
471
Image GenerationCIFAR-10--
178
Unconditional Image GenerationCIFAR-10
FID27.5
171
Out-of-Distribution DetectionCIFAR100 (test)
AUROC73
57
Out-of-Distribution DetectionSVHN (test)
AUROC0.83
48
Out-of-Distribution DetectionCelebA (test)
AUROC33
36
Out-of-Distribution DetectionCIFAR-10 Interp.
AUROC0.86
35
Image GenerationCIFAR-100 (test)
IS8.25
35
Image GenerationStacked MNIST
Modes989
32
Showing 10 of 17 rows

Other info

Follow for update