Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Penguin-VL: Exploring the Efficiency Limits of VLM with LLM-based Vision Encoders

About

Vision Language Model (VLM) development has largely relied on scaling model size, which hinders deployment on compute-constrained mobile and edge devices such as smartphones and robots. In this work, we explore the performance limits of compact (e.g., 2B and 8B) VLMs. We challenge the prevailing practice that state-of-the-art VLMs must rely on vision encoders initialized via massive contrastive pretraining (e.g., CLIP/SigLIP). We identify an objective mismatch: contrastive learning, optimized for discrimination, enforces coarse and category-level invariances that suppress fine-grained visual cues needed for dense captioning and complex VLM reasoning. To address this issue, we present Penguin-VL, whose vision encoder is initialized from a text-only LLM. Our experiments reveal that Penguin-Encoder serves as a superior alternative to traditional contrastive pretraining, unlocking a higher degree of visual fidelity and data efficiency for multimodal understanding. Across various image and video benchmarks, Penguin-VL achieves performance comparable to leading VLMs (e.g., Qwen3-VL) in mathematical reasoning and surpasses them in tasks such as document understanding, visual knowledge, and multi-perspective video understanding. Notably, these gains are achieved with a lightweight architecture, demonstrating that improved visual representation rather than model scaling is the primary driver of performance. Our ablations show that Penguin-Encoder consistently outperforms contrastive-pretrained encoders, preserving fine-grained spatial and temporal cues that are critical for dense perception and complex reasoning. This makes it a strong drop-in alternative for compute-efficient VLMs and enables high performance in resource-constrained settings. Code: https://github.com/tencent-ailab/Penguin-VL

Boqiang Zhang, Lei Ke, Ruihan Yang, Qi Gao, Tianyuan Qu, Rossell Chen, Dong Yu, Leoweiliang• 2026

Related benchmarks

TaskDatasetResultRank
Video UnderstandingMVBench
Accuracy71.7
425
Video Question AnsweringActivityNet-QA
Accuracy65.2
376
Visual Question AnsweringChartQA
Accuracy90.5
371
Mathematical ReasoningMathVista
Accuracy77.4
257
Diagram UnderstandingAI2D
Accuracy86.1
247
Optical Character RecognitionOCRBench
Score852
232
Video UnderstandingVideoMME--
222
Visual Question AnsweringRealworldQA
Accuracy75.8
179
Visual Question AnsweringDocVQA
Accuracy94.1
162
Video UnderstandingEgoSchema--
158
Showing 10 of 30 rows

Other info

GitHub

Follow for update