Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Vector Quantized Diffusion Model for Text-to-Image Synthesis

About

We present the vector quantized diffusion (VQ-Diffusion) model for text-to-image generation. This method is based on a vector quantized variational autoencoder (VQ-VAE) whose latent space is modeled by a conditional variant of the recently developed Denoising Diffusion Probabilistic Model (DDPM). We find that this latent-space method is well-suited for text-to-image generation tasks because it not only eliminates the unidirectional bias with existing methods but also allows us to incorporate a mask-and-replace diffusion strategy to avoid the accumulation of errors, which is a serious problem with existing methods. Our experiments show that the VQ-Diffusion produces significantly better text-to-image generation results when compared with conventional autoregressive (AR) models with similar numbers of parameters. Compared with previous GAN-based text-to-image methods, our VQ-Diffusion can handle more complex scenes and improve the synthesized image quality by a large margin. Finally, we show that the image generation computation in our method can be made highly efficient by reparameterization. With traditional AR methods, the text-to-image generation time increases linearly with the output image resolution and hence is quite time consuming even for normal size images. The VQ-Diffusion allows us to achieve a better trade-off between quality and speed. Our experiments indicate that the VQ-Diffusion model with the reparameterization is fifteen times faster than traditional AR methods while achieving a better image quality.

Shuyang Gu, Dong Chen, Jianmin Bao, Fang Wen, Bo Zhang, Dongdong Chen, Lu Yuan, Baining Guo• 2021

Related benchmarks

TaskDatasetResultRank
Class-conditional Image GenerationImageNet 256x256--
441
Class-conditional Image GenerationImageNet
FID5.32
132
Text-to-Image GenerationMS-COCO (val)
FID19.75
112
Conditional Image GenerationImageNet-1K 256x256 (val)
gFID11.89
86
Text-to-Image GenerationMS-COCO
FID19.75
75
Unconditional Layout GenerationRico
FID7.46
55
Class-conditional Image GenerationImageNet (val)
FID11.89
54
Text-to-Image SynthesisCOCO (test)
FID13.86
38
Text-to-Image GenerationCOCO 256 x 256 2014 (val)
FID13.86
37
Conditional layout generation (Category to Size and Position)Rico
FID3.21
27
Showing 10 of 25 rows

Other info

Code

Follow for update