Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

MultiDiffusion: Fusing Diffusion Paths for Controlled Image Generation

About

Recent advances in text-to-image generation with diffusion models present transformative capabilities in image quality. However, user controllability of the generated image, and fast adaptation to new tasks still remains an open challenge, currently mostly addressed by costly and long re-training and fine-tuning or ad-hoc adaptations to specific image generation tasks. In this work, we present MultiDiffusion, a unified framework that enables versatile and controllable image generation, using a pre-trained text-to-image diffusion model, without any further training or finetuning. At the center of our approach is a new generation process, based on an optimization task that binds together multiple diffusion generation processes with a shared set of parameters or constraints. We show that MultiDiffusion can be readily applied to generate high quality and diverse images that adhere to user-provided controls, such as desired aspect ratio (e.g., panorama), and spatial guiding signals, ranging from tight segmentation masks to bounding boxes. Project webpage: https://multidiffusion.github.io

Omer Bar-Tal, Lior Yariv, Yaron Lipman, Tali Dekel• 2023

Related benchmarks

TaskDatasetResultRank
Text-to-Image SynthesisCelebA-HQ (test)
FID240.6
19
Image GenerationLAION-COCO Horizontal
FID26.35
18
Image GenerationLAION-COCO Vertical
FID19.13
18
Human Motion CompositionBABEL
PJ0.18
13
Text-to-Image GenerationLAION-COCO
FID29.98
13
Region-based text-to-image generationCOCO 2017 (val)
FID70.93
12
Layout-to-Image GenerationCOCO-Position 2014
AP6.72
12
Image BlendingCustom image blending 20 examples (test)
Masked LPIPS0.224
9
Text-to-Image GenerationDrawBench
Spatial Fidelity (Human)55.63
8
Layout-to-Image Generation300 controllable layout images
Relativity Score0.4808
8
Showing 10 of 20 rows

Other info

Follow for update