Manifold Preserving Guided Diffusion
About
Despite the recent advancements, conditional image generation still faces challenges of cost, generalizability, and the need for task-specific training. In this paper, we propose Manifold Preserving Guided Diffusion (MPGD), a training-free conditional generation framework that leverages pretrained diffusion models and off-the-shelf neural networks with minimal additional inference cost for a broad range of tasks. Specifically, we leverage the manifold hypothesis to refine the guided diffusion steps and introduce a shortcut algorithm in the process. We then propose two methods for on-manifold training-free guidance using pre-trained autoencoders and demonstrate that our shortcut inherently preserves the manifolds when applied to latent diffusion models. Our experiments show that MPGD is efficient and effective for solving a variety of conditional generation applications in low-compute settings, and can consistently offer up to 3.8x speed-ups with the same number of diffusion steps while maintaining high sample quality compared to the baselines.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Class-conditional Image Generation | ImageNet | FID239 | 132 | |
| Conditional Image Generation | CIFAR-10 | FID88 | 71 | |
| Conditional Image Generation | Fine-grained Birds | Accuracy0.6 | 8 | |
| Conditional Image Generation | CelebA-HQ Gender+Age | Accuracy68.6 | 7 | |
| Conditional Image Generation | CelebA-HQ Gender+Hair | Accuracy63.9 | 7 | |
| Gaussian deblur | 100 images (val) | PSNR28.69 | 4 | |
| Super-Resolution | 100 images (val) | PSNR27.25 | 4 | |
| Text-to-Image Generation | HPD v2 | Rew1.0289 | 4 | |
| Text-to-Image Generation | HPD v2 | Rew1.15 | 4 |