Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Denoising Diffusion Implicit Models

About

Denoising diffusion probabilistic models (DDPMs) have achieved high quality image generation without adversarial training, yet they require simulating a Markov chain for many steps to produce a sample. To accelerate sampling, we present denoising diffusion implicit models (DDIMs), a more efficient class of iterative implicit probabilistic models with the same training procedure as DDPMs. In DDPMs, the generative process is defined as the reverse of a Markovian diffusion process. We construct a class of non-Markovian diffusion processes that lead to the same training objective, but whose reverse process can be much faster to sample from. We empirically demonstrate that DDIMs can produce high quality samples $10 \times$ to $50 \times$ faster in terms of wall-clock time compared to DDPMs, allow us to trade off computation for sample quality, and can perform semantically meaningful image interpolation directly in the latent space.

Jiaming Song, Chenlin Meng, Stefano Ermon• 2020

Related benchmarks

TaskDatasetResultRank
Image GenerationCIFAR-10 (test)
FID4.16
471
Image GenerationImageNet 256x256 (val)
FID2.12
307
Image GenerationImageNet 256x256
FID8.56
243
Unconditional Image GenerationCIFAR-10 (test)
FID4.16
216
Image GenerationCelebA 64 x 64 (test)
FID7.78
203
Image GenerationImageNet 512x512 (val)
FID-50K2.99
184
Unconditional Image GenerationCIFAR-10
FID2.2
171
Unconditional Image GenerationCIFAR-10 unconditional
FID4.04
159
Image GenerationCIFAR10 32x32 (test)
FID4.42
154
Image GenerationImageNet 64x64 resolution (test)
FID13.7
150
Showing 10 of 193 rows
...

Other info

Follow for update