Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Preventing Posterior Collapse with delta-VAEs

About

Due to the phenomenon of "posterior collapse," current latent variable generative models pose a challenging design choice that either weakens the capacity of the decoder or requires augmenting the objective so it does not only maximize the likelihood of the data. In this paper, we propose an alternative that utilizes the most powerful generative models as decoders, whilst optimising the variational lower bound all while ensuring that the latent variables preserve and encode useful information. Our proposed $\delta$-VAEs achieve this by constraining the variational family for the posterior to have a minimum distance to the prior. For sequential latent variable models, our approach resembles the classic representation learning approach of slow feature analysis. We demonstrate the efficacy of our approach at modeling text on LM1B and modeling images: learning representations, improving sample quality, and achieving state of the art log-likelihood on CIFAR-10 and ImageNet $32\times 32$.

Ali Razavi, A\"aron van den Oord, Ben Poole, Oriol Vinyals• 2019

Related benchmarks

TaskDatasetResultRank
Image GenerationCIFAR-10 (test)--
471
Density EstimationCIFAR-10 (test)
Bits/dim2.83
134
Generative ModelingCIFAR-10 (test)
NLL (bits/dim)2.83
62
Generative ModelingCIFAR-10
BPD2.83
46
Density EstimationCIFAR-10
bpd2.83
40
Generative ModelingImageNet 32x32 downsampled
Bits Per Dimension3.77
24
Language ModelingYahoo
Prior LL-330.5
18
Generative ModelingImageNet 32x32 (test)
NLL3.77
12
Density EstimationImageNet 32 x 32
NLL (bits/dim)3.77
12
Density EstimationImageNet 32x32 downsampled (val)
NLL (bits/dim)3.77
4
Showing 10 of 10 rows

Other info

Follow for update