Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

QLIP: Text-Aligned Visual Tokenization Unifies Auto-Regressive Multimodal Understanding and Generation

About

We introduce Quantized Language-Image Pretraining (QLIP), a visual tokenization method that combines state-of-the-art reconstruction quality with state-of-the-art zero-shot image understanding. QLIP trains a binary-spherical-quantization-based autoencoder with reconstruction and language-image alignment objectives. We are the first to show that the two objectives do not need to be at odds. We balance the two loss terms dynamically during training and show that a two-stage training pipeline effectively mixes the large-batch requirements of image-language pre-training with the memory bottleneck imposed by the reconstruction objective. We validate the effectiveness of QLIP for multimodal understanding and text-conditioned image generation with a single model. Specifically, QLIP serves as a drop-in replacement for the visual encoder for LLaVA and the image tokenizer for LlamaGen with comparable or even better performance. Finally, we demonstrate that QLIP enables a unified mixed-modality auto-regressive model for understanding and generation.

Yue Zhao, Fuzhao Xue, Scott Reed, Linxi Fan, Yuke Zhu, Jan Kautz, Zhiding Yu, Philipp Kr\"ahenb\"uhl, De-An Huang• 2025

Related benchmarks

TaskDatasetResultRank
Object Hallucination EvaluationPOPE--
1455
Visual Question AnsweringGQA
Accuracy61.8
1249
Multi-discipline Multimodal UnderstandingMMMU--
317
Multimodal UnderstandingMME--
207
Visual Question AnsweringGQA
Score61.8
193
Visual UnderstandingMM-Vet
MM-Vet Score33.3
142
Text-to-Image GenerationDPG-Bench
DPG Score78.17
131
Text-to-Image GenerationGenEval
GenEval Score0.48
88
Image ReconstructionImageNet-1k 256 x 256 (val)
rFID3.21
77
Visual UnderstandingMME
MME Score1.50e+3
54
Showing 10 of 13 rows

Other info

Follow for update