Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

One-step Diffusion with Distribution Matching Distillation

About

Diffusion models generate high-quality images but require dozens of forward passes. We introduce Distribution Matching Distillation (DMD), a procedure to transform a diffusion model into a one-step image generator with minimal impact on image quality. We enforce the one-step image generator match the diffusion model at distribution level, by minimizing an approximate KL divergence whose gradient can be expressed as the difference between 2 score functions, one of the target distribution and the other of the synthetic distribution being produced by our one-step generator. The score functions are parameterized as two diffusion models trained separately on each distribution. Combined with a simple regression loss matching the large-scale structure of the multi-step diffusion outputs, our method outperforms all published few-step diffusion approaches, reaching 2.62 FID on ImageNet 64x64 and 11.49 FID on zero-shot COCO-30k, comparable to Stable Diffusion but orders of magnitude faster. Utilizing FP16 inference, our model generates images at 20 FPS on modern hardware.

Tianwei Yin, Micha\"el Gharbi, Richard Zhang, Eli Shechtman, Fredo Durand, William T. Freeman, Taesung Park• 2023

Related benchmarks

TaskDatasetResultRank
Text-to-Image GenerationGenEval
GenEval Score59
360
Unconditional Image GenerationCIFAR-10
FID3.77
240
Unconditional Image GenerationCIFAR-10 (test)
FID2.62
223
Class-conditional Image GenerationImageNet 256x256 (test)
FID1.71
208
Image GenerationCIFAR-10
FID2.66
203
Unconditional Image GenerationCIFAR-10 unconditional
FID3.77
165
Class-conditional Image GenerationImageNet 64x64
FID2.62
156
Text-to-Image GenerationMJHQ-30K
Overall FID8.33
153
Image GenerationImageNet 64x64 resolution (test)
FID2.62
150
Unconditional Image GenerationCIFAR-10 32x32 (test)
FID3.77
137
Showing 10 of 35 rows

Other info

Code

Follow for update