Wasserstein Auto-Encoders
About
We propose the Wasserstein Auto-Encoder (WAE)---a new algorithm for building a generative model of the data distribution. WAE minimizes a penalized form of the Wasserstein distance between the model distribution and the target distribution, which leads to a different regularizer than the one used by the Variational Auto-Encoder (VAE). This regularizer encourages the encoded training distribution to match the prior. We compare our algorithm with several other techniques and show that it is a generalization of adversarial auto-encoders (AAE). Our experiments show that WAE shares many of the properties of VAEs (stable training, encoder-decoder architecture, nice latent manifold structure) while generating samples of better quality, as measured by the FID score.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Generation | CIFAR-10 (test) | FID93.1 | 471 | |
| Image Generation | CelebA 64 x 64 (test) | FID35 | 203 | |
| Image Generation | CelebA | FID42 | 110 | |
| Image Generation | CelebA (test) | FID66.5 | 49 | |
| Generative Modeling | MNIST (test) | -- | 35 | |
| Image Generation | Fashion (test) | FID31.5 | 16 | |
| Image Generation | SVHN latent dimension 16 (test) | FID49.07 | 13 | |
| Image Generation | CELEBA latent dimension 64 (test) | FID54.56 | 13 | |
| Image Generation | CIFAR 10 latent dimension 32 (test) | FID133 | 13 | |
| Image Generation | MNIST latent dimension 16 (test) | FID20.71 | 13 |