Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

DisDiff: Unsupervised Disentanglement of Diffusion Probabilistic Models

About

Targeting to understand the underlying explainable factors behind observations and modeling the conditional generation process on these factors, we connect disentangled representation learning to Diffusion Probabilistic Models (DPMs) to take advantage of the remarkable modeling ability of DPMs. We propose a new task, disentanglement of (DPMs): given a pre-trained DPM, without any annotations of the factors, the task is to automatically discover the inherent factors behind the observations and disentangle the gradient fields of DPM into sub-gradient fields, each conditioned on the representation of each discovered factor. With disentangled DPMs, those inherent factors can be automatically discovered, explicitly represented, and clearly injected into the diffusion process via the sub-gradient fields. To tackle this task, we devise an unsupervised approach named DisDiff, achieving disentangled representation learning in the framework of DPMs. Extensive experiments on synthetic and real-world datasets demonstrate the effectiveness of DisDiff.

Tao Yang, Yuwang Wang, Yan Lv, Nanning Zheng• 2023

Related benchmarks

TaskDatasetResultRank
Image GenerationCelebA 64 x 64 (test)
FID18.2
203
Image GenerationCelebA (test)
FID18.2
49
Disentangled Representation LearningCars3D
FactorVAE0.976
35
Disentangled Representation LearningMPI3D
FactorVAE Score0.617
18
Disentangled Representation LearningShapes3D
FactorVAE Score0.902
18
DisentanglementShapes3D
D0.723
18
DisentanglementMPI3D
D0.337
18
DisentanglementShapes3D (test)
FactorVAE0.902
13
DisentanglementCars3D
FVAE0.976
10
Disentangled Representation LearningCelebA 64x64 (test)
TAD0.305
10
Showing 10 of 12 rows

Other info

Follow for update