Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Fit and Prune: Fast and Training-free Visual Token Pruning for Multi-modal Large Language Models

About

Recent progress in Multimodal Large Language Models(MLLMs) often use large image tokens to compensate the visual shortcoming of MLLMs, which not only exhibits obvious redundancy but also greatly exacerbates the already high computation. Token pruning is an effective solution for speeding up MLLMs, but when and how to drop tokens still remains a challenge. In this paper, we propose a novel and training-free approach for the effective visual token pruning of MLLMs, termed FitPrune, which can quickly produce a complete pruning recipe for MLLMs according to a pre-defined budget. Specifically, FitPrune considers token pruning as a statistical problem of MLLM and its objective is to find out an optimal pruning scheme that can minimize the divergence of the attention distributions before and after pruning. In practice, FitPrune can be quickly accomplished based on the attention statistics from a small batch of inference data, avoiding the expensive trials of MLLMs. According to the pruning recipe, an MLLM can directly remove the redundant visual tokens of different examples during inference. To validate FitPrune, we apply it to a set of recent MLLMs, including LLaVA-1.5, LLaVA-HR and LLaVA-NEXT, and conduct extensive experiments on a set of benchmarks. The experimental results show that our FitPrune can not only reduce the computational complexity to a large extent, while retaining high performance, e.g., -54.9% FLOPs for LLaVA-NEXT with only 0.5% accuracy drop. Notably, the pruning recipe can be obtained in about 5 minutes. Our code is available at https://github.com/ywh187/FitPrune.

Weihao Ye, Qiong Wu, Wenhao Lin, Yiyi Zhou• 2024

Related benchmarks

TaskDatasetResultRank
Object Hallucination EvaluationPOPE
Accuracy53.8
1455
Visual Question AnsweringTextVQA
Accuracy57.4
1285
Visual Question AnsweringVQA v2 (test-dev)
Overall Accuracy62.7
706
Multimodal UnderstandingMMBench--
637
Science Question AnsweringScienceQA IMG
Accuracy68.3
294
Multimodal EvaluationMM-Vet
Score21.8
180
Multimodal BenchmarkingMMBench English
Accuracy56.2
125
Multimodal Model EvaluationMME
Score1.83e+3
98
Visual Question AnsweringGQA
GQA Score60.4
85
Multi-modal EvaluationMME (total)
MME Total Score1.15e+3
76
Showing 10 of 15 rows

Other info

Follow for update