Analytic-DPM: an Analytic Estimate of the Optimal Reverse Variance in Diffusion Probabilistic Models
About
Diffusion probabilistic models (DPMs) represent a class of powerful generative models. Despite their success, the inference of DPMs is expensive since it generally needs to iterate over thousands of timesteps. A key problem in the inference is to estimate the variance in each timestep of the reverse process. In this work, we present a surprising result that both the optimal reverse variance and the corresponding optimal KL divergence of a DPM have analytic forms w.r.t. its score function. Building upon it, we propose Analytic-DPM, a training-free inference framework that estimates the analytic forms of the variance and KL divergence using the Monte Carlo method and a pretrained score-based model. Further, to correct the potential bias caused by the score-based model, we derive both lower and upper bounds of the optimal variance and clip the estimate for a better result. Empirically, our analytic-DPM improves the log-likelihood of various DPMs, produces high-quality samples, and meanwhile enjoys a 20x to 80x speed up.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Unconditional Image Generation | CIFAR-10 (test) | FID5.71 | 216 | |
| Image Generation | CelebA 64 x 64 (test) | FID12.25 | 203 | |
| Unconditional Image Generation | CIFAR-10 | FID3.04 | 171 | |
| Image Generation | CIFAR10 32x32 (test) | FID8.65 | 154 | |
| Image Generation | ImageNet 64x64 resolution (test) | FID41.56 | 150 | |
| Unconditional Image Generation | CelebA unconditional 64 x 64 | FID3.13 | 95 | |
| Image Generation | CIFAR-10 discrete-time and continuous-time models (test) | FID3.6 | 92 | |
| Unconditional Image Generation | FFHQ 256x256 | FID5.98 | 64 |