Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Emerging Properties in Unified Multimodal Pretraining

About

Unifying multimodal understanding and generation has shown impressive capabilities in cutting-edge proprietary systems. In this work, we introduce BAGEL, an open-source foundational model that natively supports multimodal understanding and generation. BAGEL is a unified, decoder-only model pretrained on trillions of tokens curated from large-scale interleaved text, image, video, and web data. When scaled with such diverse multimodal interleaved data, BAGEL exhibits emerging capabilities in complex multimodal reasoning. As a result, it significantly outperforms open-source unified models in both multimodal generation and understanding across standard benchmarks, while exhibiting advanced multimodal reasoning abilities such as free-form image manipulation, future frame prediction, 3D manipulation, and world navigation. In the hope of facilitating further opportunities for multimodal research, we share the key findings, pretraining details, data creation protocal, and release our code and checkpoints to the community. The project page is at https://bagel-ai.org/

Chaorui Deng, Deyao Zhu, Kunchang Li, Chenhui Gou, Feng Li, Zeyu Wang, Shu Zhong, Weihao Yu, Xiaonan Nie, Ziang Song, Guang Shi, Haoqi Fan• 2025

Related benchmarks

TaskDatasetResultRank
Text-based Visual Question AnsweringTextVQA
Accuracy80
496
Text-to-Image GenerationGenEval
Overall Score88
467
Multimodal UnderstandingMM-Vet
MM-Vet Score67.2
418
Multimodal UnderstandingMMBench--
367
Mathematical ReasoningMathVista
Score73.1
322
OCR EvaluationOCRBench
Score78.3
296
Multimodal ReasoningMM-Vet
MM-Vet Score67.2
281
Text-to-Image GenerationGenEval
GenEval Score88
277
Multi-discipline Multimodal UnderstandingMMMU--
266
Mathematical ReasoningAIME 2024
Accuracy10
251
Showing 10 of 264 rows
...

Other info

Code

Follow for update