Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Stop Looking for Important Tokens in Multimodal Language Models: Duplication Matters More

About

Vision tokens in multimodal large language models often dominate huge computational overhead due to their excessive length compared to linguistic modality. Abundant recent methods aim to solve this problem with token pruning, which first defines an importance criterion for tokens and then prunes the unimportant vision tokens during inference. However, in this paper, we show that the importance is not an ideal indicator to decide whether a token should be pruned. Surprisingly, it usually results in inferior performance than random token pruning and leading to incompatibility to efficient attention computation operators.Instead, we propose DART (Duplication-Aware Reduction of Tokens), which prunes tokens based on its duplication with other tokens, leading to significant and training-free acceleration. Concretely, DART selects a small subset of pivot tokens and then retains the tokens with low duplication to the pivots, ensuring minimal information loss during token pruning. Experiments demonstrate that DART can prune 88.9% vision tokens while maintaining comparable performance, leading to a 1.99$\times$ and 2.99$\times$ speed-up in total time and prefilling stage, respectively, with good compatibility to efficient attention operators. Our codes are available at https://github.com/ZichenWen1/DART.

Zichen Wen, Yifeng Gao, Shaobo Wang, Junyuan Zhang, Qintong Zhang, Weijia Li, Conghui He, Linfeng Zhang• 2025

Related benchmarks

TaskDatasetResultRank
Visual Question AnsweringVizWiz
Accuracy98.57
1525
Object Hallucination EvaluationPOPE
Accuracy89.2
1455
Visual Question AnsweringVQA v2
Accuracy79.47
1362
Visual Question AnsweringTextVQA
Accuracy82.1
1285
Visual Question AnsweringGQA
Accuracy61.7
1249
Text-based Visual Question AnsweringTextVQA
Accuracy70.4
807
Visual Question AnsweringVQA v2 (test-dev)
Overall Accuracy75.7
706
Multimodal EvaluationMME
Score2.25e+3
658
Multimodal UnderstandingMMBench
Accuracy79.6
637
Visual Question AnsweringGQA
Accuracy61.7
505
Showing 10 of 141 rows
...

Other info

Follow for update