Cold Diffusion: Inverting Arbitrary Image Transforms Without Noise
About
Standard diffusion models involve an image transform -- adding Gaussian noise -- and an image restoration operator that inverts this degradation. We observe that the generative behavior of diffusion models is not strongly dependent on the choice of image degradation, and in fact an entire family of generative models can be constructed by varying this choice. Even when using completely deterministic degradations (e.g., blur, masking, and more), the training and test-time update rules that underlie diffusion models can be easily generalized to create generative models. The success of these fully deterministic models calls into question the community's understanding of diffusion models, which relies on noise in either gradient Langevin dynamics or variational inference, and paves the way for generalized diffusion models that invert arbitrary processes. Our code is available at https://github.com/arpitbansal297/Cold-Diffusion-Models
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Unconditional Image Generation | CelebA unconditional 64 x 64 | FID7.13 | 95 | |
| Unconditional Image Generation | FFHQ 256x256 | FID28.11 | 64 | |
| Image Reconstruction | CIFAR-10 | -- | 25 | |
| Image Reconstruction | CIFAR-10 (test) | SSIM0.942 | 15 | |
| Image Reconstruction | CelebA (test) | -- | 15 | |
| Image Generation | CelebA 128x128 | FID49.45 | 4 | |
| Image Generation | AFHQ 128x128 | FID54.68 | 4 |