Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Unsupervised Discovery of Interpretable Directions in the GAN Latent Space

About

The latent spaces of GAN models often have semantically meaningful directions. Moving in these directions corresponds to human-interpretable image transformations, such as zooming or recoloring, enabling a more controllable generation process. However, the discovery of such directions is currently performed in a supervised manner, requiring human labels, pretrained models, or some form of self-supervision. These requirements severely restrict a range of directions existing approaches can discover. In this paper, we introduce an unsupervised method to identify interpretable directions in the latent space of a pretrained GAN model. By a simple model-agnostic procedure, we find directions corresponding to sensible semantic manipulations without any form of (self-)supervision. Furthermore, we reveal several non-trivial findings, which would be difficult to obtain by existing methods, e.g., a direction corresponding to background removal. As an immediate practical benefit of our work, we show how to exploit this finding to achieve competitive performance for weakly-supervised saliency detection.

Andrey Voynov, Artem Babenko• 2020

Related benchmarks

TaskDatasetResultRank
Disentangled Representation LearningCars3D
FactorVAE0.852
35
Disentangled Representation LearningMPI3D
FactorVAE Score0.391
18
Disentangled Representation LearningShapes3D
FactorVAE Score0.805
18
DisentanglementShapes3D
D0.38
18
DisentanglementMPI3D
D0.196
18
Attribute DisentanglementCelebA-HQ
Age Attribute Score175
15
DisentanglementCars3D
FVAE0.852
10
Missing Attribute DiscoveryCelebA 7 controlled experiments
Ascore0.544
6
Attribute EditingCelebA-HQ
Age Score26.1
3
Showing 9 of 9 rows

Other info

Follow for update