Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Hierarchical Text-Conditional Image Generation with CLIP Latents

About

Contrastive models like CLIP have been shown to learn robust representations of images that capture both semantics and style. To leverage these representations for image generation, we propose a two-stage model: a prior that generates a CLIP image embedding given a text caption, and a decoder that generates an image conditioned on the image embedding. We show that explicitly generating image representations improves image diversity with minimal loss in photorealism and caption similarity. Our decoders conditioned on image representations can also produce variations of an image that preserve both its semantics and style, while varying the non-essential details absent from the image representation. Moreover, the joint embedding space of CLIP enables language-guided image manipulations in a zero-shot fashion. We use diffusion models for the decoder and experiment with both autoregressive and diffusion models for the prior, finding that the latter are computationally more efficient and produce higher-quality samples.

Aditya Ramesh, Prafulla Dhariwal, Alex Nichol, Casey Chu, Mark Chen• 2022

Related benchmarks

TaskDatasetResultRank
Image ClassificationDTD
Accuracy34.5
487
Text-to-Image GenerationGenEval
Overall Score52
467
ClassificationCars
Accuracy48.3
314
Text-to-Image GenerationGenEval
GenEval Score52
277
Image ClassificationPets
Accuracy61.7
204
Text-to-Image GenerationDPG-Bench
Overall Score68.09
173
Text-to-Image GenerationGenEval (test)
Two Obj. Acc66
169
Image ClassificationCaltech101
Base Accuracy61.3
129
Text-to-Image GenerationMS-COCO 2014 (val)
FID10.39
128
Image ClassificationFlowers
Accuracy84.1
127
Showing 10 of 87 rows
...

Other info

Follow for update