Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

BREEN: Bridge Data-Efficient Encoder-Free Multimodal Learning with Learnable Queries

About

Encoder-free multimodal large language models(MLLMs) eliminate the need for a well-trained vision encoder by directly processing image tokens before the language model. While this approach reduces computational overhead and model complexity, it often requires large amounts of training data to effectively capture the visual knowledge typically encoded by vision models like CLIP. The absence of a vision encoder implies that the model is likely to rely on substantial data to learn the necessary visual-semantic alignments. In this work, we present BREEN, a data-efficient encoder-free multimodal architecture that mitigates this issue. BREEN leverages a learnable query and image experts to achieve comparable performance with significantly less training data. The learnable query, positioned between image and text tokens, is supervised by the output of a pretrained CLIP model to distill visual knowledge, bridging the gap between visual and textual modalities. Additionally, the image expert processes image tokens and learnable queries independently, improving efficiency and reducing interference with the LLM's textual capabilities. BREEN achieves comparable performance to prior encoder-free state-of-the-art models like Mono-InternVL, using only 13 million text-image pairs in training about one percent of the data required by existing methods. Our work highlights a promising direction for data-efficient encoder-free multimodal learning, offering an alternative to traditional encoder-based approaches.

Tianle Li, Yongming Rao, Winston Hu, Yu Cheng• 2025

Related benchmarks

TaskDatasetResultRank
Text-based Visual Question AnsweringTextVQA
Accuracy65.7
496
Multi-discipline Multimodal UnderstandingMMMU--
266
Visual Question AnsweringAI2D
Accuracy76.4
174
Multi-modal Vision-Language UnderstandingMMVet
Score38.9
38
General Vision-Language UnderstandingMMB
Score71.4
25
Hallucination EvaluationHallB
Score37
19
High-quality Vision-Language EvaluationMMStar
Score51.2
14
Showing 7 of 7 rows

Other info

Follow for update