Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Manifold Preserving Guided Diffusion

About

Despite the recent advancements, conditional image generation still faces challenges of cost, generalizability, and the need for task-specific training. In this paper, we propose Manifold Preserving Guided Diffusion (MPGD), a training-free conditional generation framework that leverages pretrained diffusion models and off-the-shelf neural networks with minimal additional inference cost for a broad range of tasks. Specifically, we leverage the manifold hypothesis to refine the guided diffusion steps and introduce a shortcut algorithm in the process. We then propose two methods for on-manifold training-free guidance using pre-trained autoencoders and demonstrate that our shortcut inherently preserves the manifolds when applied to latent diffusion models. Our experiments show that MPGD is efficient and effective for solving a variety of conditional generation applications in low-compute settings, and can consistently offer up to 3.8x speed-ups with the same number of diffusion steps while maintaining high sample quality compared to the baselines.

Yutong He, Naoki Murata, Chieh-Hsin Lai, Yuhta Takida, Toshimitsu Uesaka, Dongjun Kim, Wei-Hsiang Liao, Yuki Mitsufuji, J. Zico Kolter, Ruslan Salakhutdinov, Stefano Ermon• 2023

Related benchmarks

TaskDatasetResultRank
Class-conditional Image GenerationImageNet
FID239
132
Conditional Image GenerationCIFAR-10
FID88
71
Conditional Image GenerationFine-grained Birds
Accuracy0.6
8
Conditional Image GenerationCelebA-HQ Gender+Age
Accuracy68.6
7
Conditional Image GenerationCelebA-HQ Gender+Hair
Accuracy63.9
7
Gaussian deblur100 images (val)
PSNR28.69
4
Super-Resolution100 images (val)
PSNR27.25
4
Text-to-Image GenerationHPD v2
Rew1.0289
4
Text-to-Image GenerationHPD v2
Rew1.15
4
Showing 9 of 9 rows

Other info

Follow for update