Unsupervised Discovery of Interpretable Directions in the GAN Latent Space
About
The latent spaces of GAN models often have semantically meaningful directions. Moving in these directions corresponds to human-interpretable image transformations, such as zooming or recoloring, enabling a more controllable generation process. However, the discovery of such directions is currently performed in a supervised manner, requiring human labels, pretrained models, or some form of self-supervision. These requirements severely restrict a range of directions existing approaches can discover. In this paper, we introduce an unsupervised method to identify interpretable directions in the latent space of a pretrained GAN model. By a simple model-agnostic procedure, we find directions corresponding to sensible semantic manipulations without any form of (self-)supervision. Furthermore, we reveal several non-trivial findings, which would be difficult to obtain by existing methods, e.g., a direction corresponding to background removal. As an immediate practical benefit of our work, we show how to exploit this finding to achieve competitive performance for weakly-supervised saliency detection.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Disentangled Representation Learning | Cars3D | FactorVAE0.852 | 35 | |
| Disentangled Representation Learning | MPI3D | FactorVAE Score0.391 | 18 | |
| Disentangled Representation Learning | Shapes3D | FactorVAE Score0.805 | 18 | |
| Disentanglement | Shapes3D | D0.38 | 18 | |
| Disentanglement | MPI3D | D0.196 | 18 | |
| Attribute Disentanglement | CelebA-HQ | Age Attribute Score175 | 15 | |
| Disentanglement | Cars3D | FVAE0.852 | 10 | |
| Missing Attribute Discovery | CelebA 7 controlled experiments | Ascore0.544 | 6 | |
| Attribute Editing | CelebA-HQ | Age Score26.1 | 3 |