Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

VScan: Rethinking Visual Token Reduction for Efficient Large Vision-Language Models

About

Recent Large Vision-Language Models (LVLMs) have advanced multi-modal understanding by incorporating finer-grained visual perception and encoding. However, such methods incur significant computational costs due to longer visual token sequences, posing challenges for real-time deployment. To mitigate this, prior studies have explored pruning unimportant visual tokens either at the output layer of the visual encoder or at the early layers of the language model. In this work, we revisit these design choices and reassess their effectiveness through comprehensive empirical studies of how visual tokens are processed throughout the visual encoding and language decoding stages. Guided by these insights, we propose VScan, a two-stage visual token reduction framework that addresses token redundancy by: (1) integrating complementary global and local scans with token merging during visual encoding, and (2) introducing pruning at intermediate layers of the language model. Extensive experimental results across four LVLMs validate the effectiveness of VScan in accelerating inference and demonstrate its superior performance over current state-of-the-arts on sixteen benchmarks. Notably, when applied to LLaVA-NeXT-7B, VScan achieves a 2.91$\times$ speedup in prefilling and a 10$\times$ reduction in FLOPs, while retaining 95.4\% of the original performance. Code is available at https://github.com/Tencent/SelfEvolvingAgent/tree/main/VScan.

Ce Zhang, Kaixin Ma, Tianqing Fang, Wenhao Yu, Hongming Zhang, Zhisong Zhang, Haitao Mi, Dong Yu• 2025

Related benchmarks

TaskDatasetResultRank
Visual Question AnsweringVQA v2
Accuracy75.4
1165
Object Hallucination EvaluationPOPE
Accuracy85
935
Text-based Visual Question AnsweringTextVQA
Accuracy55.6
496
Visual Question AnsweringGQA
Accuracy58.3
374
Visual GroundingRefCOCO+ (val)
Accuracy79
171
Visual GroundingRefCOCO+ (testB)
Accuracy70.6
169
Visual GroundingRefCOCO+ (testA)
Accuracy84.6
168
Multimodal UnderstandingMMBench CN
Accuracy55.7
162
Visual GroundingRefCOCO (testB)
Accuracy82.2
125
Visual GroundingRefCOCO (val)
Accuracy86.7
119
Showing 10 of 17 rows

Other info

Follow for update