Topological Autoencoders
About
We propose a novel approach for preserving topological structures of the input space in latent representations of autoencoders. Using persistent homology, a technique from topological data analysis, we calculate topological signatures of both the input and latent space to derive a topological loss term. Under weak theoretical assumptions, we construct this loss in a differentiable manner, such that the encoding learns to retain multi-scale connectivity information. We show that our approach is theoretically well-founded and that it exhibits favourable latent representations on a synthetic manifold as well as on real-world image data sets, while preserving low reconstruction errors.
Michael Moor, Max Horn, Bastian Rieck, Karsten Borgwardt• 2019
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Classification | warpPIE 10P | Accuracy73 | 26 | |
| Classification | PROT579 | Accuracy88.3 | 8 | |
| Classification | MC1374 | Accuracy61.3 | 8 | |
| Classification | HCL500 | Accuracy56 | 8 | |
| Classification | GA1457 | Accuracy0.746 | 8 | |
| Classification | SAM561 | Accuracy72.4 | 8 | |
| Dimensionality Reduction | SPHERES | KL Div (0.01)0.085 | 6 | |
| Dimensionality Reduction | CIFAR-10 | KL Divergence (0.01)0.556 | 5 | |
| Dimensionality Reduction | Fashion MNIST | KL Divergence (0.01)0.392 | 5 | |
| Dimensionality Reduction | MNIST | KL Divergence (0.01)0.341 | 5 |
Showing 10 of 10 rows