Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

SEED-X: Multimodal Models with Unified Multi-granularity Comprehension and Generation

About

The rapid evolution of multimodal foundation model has demonstrated significant progresses in vision-language understanding and generation, e.g., our previous work SEED-LLaMA. However, there remains a gap between its capability and the real-world applicability, primarily due to the model's limited capacity to effectively respond to various user instructions and interact with diverse visual data. In this work, we focus on bridging this gap through integrating two enhanced features: (1) comprehending images of arbitrary sizes and ratios, and (2) enabling multi-granularity image generation. We present a unified and versatile foundation model, namely, SEED-X, which is able to model multi-granularity visual semantics for comprehension and generation tasks. Besides the competitive results on public benchmarks, SEED-X demonstrates its effectiveness in handling real-world applications across various domains after instruction tuning. We hope that our work will inspire future research into what can be achieved by versatile multimodal foundation models in real-world applications. The models, codes, and datasets are released in https://github.com/AILab-CVC/SEED-X.

Yuying Ge, Sijie Zhao, Jinguo Zhu, Yixiao Ge, Kun Yi, Lin Song, Chen Li, Xiaohan Ding, Ying Shan• 2024

Related benchmarks

TaskDatasetResultRank
Object Hallucination EvaluationPOPE
Accuracy84.2
1455
Visual Question AnsweringGQA
Accuracy47.9
1249
Multimodal UnderstandingMMBench--
637
Text-to-Image GenerationGenEval
Overall Score49
506
Multimodal UnderstandingMMMU
Accuracy35.6
437
Text-to-Image GenerationGenEval
Overall Score49
391
Text-to-Image GenerationGenEval
GenEval Score49
360
Multimodal UnderstandingSEED-Bench--
343
Multi-discipline Multimodal UnderstandingMMMU--
317
Text-to-Image GenerationGenEval (test)
Two Obj. Acc58
221
Showing 10 of 49 rows

Other info

Follow for update