Adaptive Moments are Surprisingly Effective for Plug-and-Play Diffusion Sampling
About
Guided diffusion sampling relies on approximating often intractable likelihood scores, which introduces significant noise into the sampling dynamics. We propose using adaptive moment estimation to stabilize these noisy likelihood scores during sampling. Despite its simplicity, our approach achieves state-of-the-art results on image restoration and class-conditional generation tasks, outperforming more complicated methods, which are often computationally more expensive. We provide empirical analysis of our method on both synthetic and real data, demonstrating that mitigating gradient noise through adaptive moments offers an effective way to improve alignment.
Christian Belardi, Justin Lovelace, Kilian Q. Weinberger, Carla P. Gomes• 2026
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Class-conditional generation | ImageNet | FID29.6 | 14 | |
| Super-Resolution | ImageNet 4x scale | LPIPS0.12 | 14 | |
| Super-Resolution | ImageNet 16x scale | LPIPS0.27 | 14 | |
| Gaussian Deblur 12 | Cats | LPIPS0.27 | 14 | |
| Gaussian Deblurring | ImageNet Gaussian Blur sigma=12 | LPIPS0.33 | 14 | |
| Inpainting | Cats | LPIPS0.08 | 14 | |
| Super Resolution 16x | Cats | LPIPS0.24 | 14 | |
| Gaussian Deblurring | ImageNet Gaussian Blur sigma=3 | LPIPS0.17 | 14 | |
| Gaussian Deblur 3 | Cats | LPIPS0.14 | 14 | |
| Super-Resolution (4x) | Cats | LPIPS0.09 | 14 |
Showing 10 of 19 rows