Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Diagnosing and Enhancing VAE Models

About

Although variational autoencoders (VAEs) represent a widely influential deep generative model, many aspects of the underlying energy function remain poorly understood. In particular, it is commonly believed that Gaussian encoder/decoder assumptions reduce the effectiveness of VAEs in generating realistic samples. In this regard, we rigorously analyze the VAE objective, differentiating situations where this belief is and is not actually true. We then leverage the corresponding insights to develop a simple VAE enhancement that requires no additional hyperparameters or sensitive tuning. Quantitatively, this proposal produces crisp samples and stable FID scores that are actually competitive with a variety of GAN models, all while retaining desirable attributes of the original VAE architecture. A shorter version of this work will appear in the ICLR 2019 conference proceedings (Dai and Wipf, 2019). The code for our model is available at https://github.com/daib13/ TwoStageVAE.

Bin Dai, David Wipf• 2019

Related benchmarks

TaskDatasetResultRank
Image GenerationCIFAR-10 (test)
FID72.9
471
Image GenerationCelebA 64 x 64 (test)
FID44.4
203
Image GenerationCIFAR-10--
178
Image GenerationCelebA
FID44.4
110
Image GenerationCelebA (test)
FID44.4
49
Image GenerationMNIST
FID12.6
44
Image GenerationFashion MNIST
FID29.3
38
Generative ModelingMNIST (test)--
35
Image GenerationFashion (test)
FID26.1
16
Image GenerationSVHN (test)
FID42.81
14
Showing 10 of 16 rows

Other info

Follow for update