Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

IWP: Token Pruning as Implicit Weight Pruning in Large Vision Language Models

About

Large Vision Language Models show impressive performance across image and video understanding tasks, yet their computational cost grows rapidly with the number of visual tokens. Existing token pruning methods mitigate this issue through empirical approaches while overlooking the internal mechanism of attention. In this paper, we propose a novel training free token pruning framework grounded in the dual form perspective of attention. We reformulate attention as an implicit linear layer whose weight matrix is the sum of rank 1 outer products, each generated by a single token's key value pair. Token pruning thus reduces to selecting an optimal subset of these rank 1 updates that best approximates the original dual weight matrix. Extending this perspective to standard softmax attention in LVLMs, we derive a novel metric quantifying both a token's information magnitude and information duplication. To efficiently select the subset with the proposed metric, we introduce Progressive Chunked Maximal Marginal Relevance. Extensive experiments demonstrate that our method achieves a better trade off between performance and efficiency, while providing another perspective on existing pruning approaches.

Dong-Jae Lee, Sunghyun Baek, Junmo Kim• 2026

Related benchmarks

TaskDatasetResultRank
Text-based Visual Question AnsweringTextVQA
Accuracy75.7
807
Multimodal UnderstandingMMBench
Accuracy84.2
637
Multimodal UnderstandingMMMU
Accuracy58.9
437
Multimodal UnderstandingMMStar
Accuracy61
324
Document Visual Question AnsweringDocVQA
ANLS82.1
263
Video UnderstandingMLVU
Score60.9
221
Video UnderstandingEgoSchema
EgoSchema Score62.2
158
Multimodal UnderstandingLVLM Evaluation Suite (AI2D, DocVQA, InfoVQA, MMBench, MME, MMMU, SciQA, TextVQA, MMStar, POPE)
AI2D81.8
38
Scientific Question AnsweringSciQA
Accuracy94.2
35
Multimodal UnderstandingMME
Score2.38e+3
16
Showing 10 of 14 rows

Other info

Follow for update