Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Understanding disentangling in $\beta$-VAE

About

We present new intuitions and theoretical assessments of the emergence of disentangled representation in variational autoencoders. Taking a rate-distortion theory perspective, we show the circumstances under which representations aligned with the underlying generative factors of variation of data emerge when optimising the modified ELBO bound in $\beta$-VAE, as training progresses. From these insights, we propose a modification to the training regime of $\beta$-VAE, that progressively increases the information capacity of the latent code during training. This modification facilitates the robust learning of disentangled representations in $\beta$-VAE, without the previous trade-off in reconstruction accuracy.

Christopher P. Burgess, Irina Higgins, Arka Pal, Loic Matthey, Nick Watters, Guillaume Desjardins, Alexander Lerchner• 2018

Related benchmarks

TaskDatasetResultRank
Image ReconstructionCelebA-HQ (test)
FID (Reconstruction)145.3
50
DisentanglementMPI3D (test)
DCI30.95
17
DisentanglementSmallNORB (test)
DCI33.14
17
Image ClassificationCelebA-HQ (test)
F1 Score68.94
13
DisentanglementCelebA-HQ (test)
Disentanglement42.99
13
Disentanglement AnalysisCUB (full)
Disentanglement37.53
8
Disentanglement AnalysisCUB (test)
DCI (average)46.94
8
Disentanglement AnalysisMPI3D Toy
Disen.18.43
8
Disentanglement AnalysisMPI3D realistic
Disentanglement Score18.76
8
Disentanglement AnalysisMPI3D real
Disentanglement Score17.62
8
Showing 10 of 18 rows

Other info

Follow for update