Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Liquid: Language Models are Scalable and Unified Multi-modal Generators

About

We present Liquid, an auto-regressive generation paradigm that seamlessly integrates visual comprehension and generation by tokenizing images into discrete codes and learning these code embeddings alongside text tokens within a shared feature space for both vision and language. Unlike previous multimodal large language model (MLLM), Liquid achieves this integration using a single large language model (LLM), eliminating the need for external pretrained visual embeddings such as CLIP. For the first time, Liquid uncovers a scaling law that performance drop unavoidably brought by the unified training of visual and language tasks diminishes as the model size increases. Furthermore, the unified token space enables visual generation and comprehension tasks to mutually enhance each other, effectively removing the typical interference seen in earlier models. We show that existing LLMs can serve as strong foundations for Liquid, saving 100x in training costs while outperforming Chameleon in multimodal capabilities and maintaining language performance comparable to mainstream LLMs like LLAMA2. Liquid also outperforms models like SD v2.1 and SD-XL (FID of 5.47 on MJHQ-30K), excelling in both vision-language and text-only tasks. This work demonstrates that LLMs such as Qwen2.5 and GEMMA2 are powerful multimodal generators, offering a scalable solution for enhancing both vision-language understanding and generation. The code and models will be released at https://github.com/FoundationVision/Liquid.

Junfeng Wu, Yi Jiang, Chuofan Ma, Yuliang Liu, Hengshuang Zhao, Zehuan Yuan, Song Bai, Xiang Bai• 2024

Related benchmarks

TaskDatasetResultRank
Object Hallucination EvaluationPOPE--
1455
Multimodal UnderstandingMM-Vet--
531
Multimodal UnderstandingMME
MME Score1.45e+3
207
Visual Question AnsweringGQA
Score71.3
193
Text-to-Image GenerationT2I-CompBench
Shape Fidelity52.3
185
Text-to-Image GenerationDPG-Bench
DPG Score83.45
131
Multimodal UnderstandingPOPE
POPE Score0.832
90
Text-to-Image GenerationGenEval
GenEval Score0.55
88
Visual Question AnsweringVQAv2 (test)
VQA Accuracy63.5
82
Multimodal Visual PerceptionMMVP
Accuracy58.33
72
Showing 10 of 43 rows

Other info

Follow for update