On the Latent Space of Wasserstein Auto-Encoders
About
We study the role of latent space dimensionality in Wasserstein auto-encoders (WAEs). Through experimentation on synthetic and real datasets, we argue that random encoders should be preferred over deterministic encoders. We highlight the potential of WAEs for representation learning with promising results on a benchmark disentanglement task.
Paul K. Rubenstein, Bernhard Schoelkopf, Ilya Tolstikhin• 2018
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Sentence Interpolation Smoothness | ARGO randomly sampled 200 sentence pairs | Average IS0.172 | 22 | |
| Autoencoding | Mathematical expressions EVAL (test) | BLEU41 | 22 | |
| Language modelling | Explanatory sentences | BLEU26 | 19 | |
| Language modelling | Mathematical expression EVAL (test) | Exact Match0.00e+0 | 19 | |
| Autoencoding | Explanatory sentences (test) | BLEU26 | 13 | |
| Autoencoding | Mathematical expressions EASY (test) | BLEU Score39 | 11 | |
| Autoencoding | Mathematical expressions VAR (test) | BLEU45 | 11 | |
| Autoencoding | Mathematical expressions LEN (test) | BLEU0.49 | 11 | |
| Explanatory Inference Retrieval | WorldTree 1.0 (test) | MAP (t=1)19.13 | 7 |
Showing 9 of 9 rows