SenseFlow: Scaling Distribution Matching for Flow-based Text-to-Image Distillation
About
The Distribution Matching Distillation (DMD) has been successfully applied to text-to-image diffusion models such as Stable Diffusion (SD) 1.5. However, vanilla DMD suffers from convergence difficulties on large-scale flow-based text-to-image models, such as SD 3.5 and FLUX. In this paper, we first analyze the issues when applying vanilla DMD on large-scale models. Then, to overcome the scalability challenge, we propose implicit distribution alignment (IDA) to regularize the distance between the generator and fake distribution. Furthermore, we propose intra-segment guidance (ISG) to relocate the timestep importance distribution from the teacher model. With IDA alone, DMD converges for SD 3.5; employing both IDA and ISG, DMD converges for SD 3.5 and FLUX.1 dev. Along with other improvements such as scaled up discriminator models, our final model, dubbed \textbf{SenseFlow}, achieves superior performance in distillation for both diffusion based text-to-image models such as SDXL, and flow-matching models such as SD 3.5 Large and FLUX. The source code will be avaliable at https://github.com/XingtongGe/SenseFlow.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Text-to-Image Generation | GenEval | GenEval Score60 | 277 | |
| Text-to-Image Generation | DPG-Bench | DPG Score79.86 | 89 | |
| Text-to-Image Generation | OneIG-Bench | Alignment0.776 | 33 | |
| Text-to-Image Generation | MS-COCO 10K prompts 2014 (val) | FID34.1 | 19 | |
| Text-to-Image Generation | HPS prompt set v2 | CLIP Score0.283 | 11 | |
| Text-to-Image Generation | Align5000 1.0 (test) | CLIP Score0.311 | 9 |