Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

FlowCut: Rethinking Redundancy via Information Flow for Efficient Vision-Language Models

About

Large vision-language models (LVLMs) excel at multimodal understanding but suffer from high computational costs due to redundant vision tokens. Existing pruning methods typically rely on single-layer attention scores to rank and prune redundant visual tokens to solve this inefficiency. However, as the interaction between tokens and layers is complicated, this raises a basic question: Is such a simple single-layer criterion sufficient to identify redundancy? To answer this question, we rethink the emergence of redundant visual tokens from a fundamental perspective: information flow, which models the interaction between tokens and layers by capturing how information moves between tokens across layers. We find (1) the CLS token acts as an information relay, which can simplify the complicated flow analysis; (2) the redundancy emerges progressively and dynamically via layer-wise attention concentration; and (3) relying solely on attention scores from single layers can lead to contradictory redundancy identification. Based on this, we propose FlowCut, an information-flow-aware pruning framework, mitigating the insufficiency of the current criterion for identifying redundant tokens and better aligning with the model's inherent behaviors. Extensive experiments show that FlowCut achieves superior results, outperforming SoTA by 1.6% on LLaVA-1.5-7B with 88.9% token reduction, and by 4.3% on LLaVA-NeXT-7B with 94.4% reduction, delivering 3.2x speed-up in the prefilling stage. Our code is available at https://github.com/TungChintao/FlowCut

Jintao Tong, Wenwei Jin, Pengda Qin, Anqi Li, Yixiong Zou, Yuhong Li, Yuhua Li, Ruixuan Li• 2025

Related benchmarks

TaskDatasetResultRank
Visual Question AnsweringVQA v2
Accuracy72.8
1165
Object Hallucination EvaluationPOPE
Accuracy80.2
935
Text-based Visual Question AnsweringTextVQA
Accuracy55.6
496
Visual Question AnsweringGQA
Accuracy55.6
374
Multimodal UnderstandingMMBench CN
Accuracy55.4
162
Science Question AnsweringScienceQA SQA-IMG
Accuracy69.1
114
Multimodal UnderstandingMMBench (MMB)
Accuracy60.8
69
Visual Question AnsweringVQA v2
Accuracy (Clean)72.9
37
Visual Question AnsweringTextVQA
Clean Accuracy57.8
37
Visual Question AnsweringGQA
Clean Accuracy55.4
27
Showing 10 of 10 rows

Other info

Follow for update