Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

ApET: Approximation-Error Guided Token Compression for Efficient VLMs

About

Recent Vision-Language Models (VLMs) have demonstrated remarkable multimodal understanding capabilities, yet the redundant visual tokens incur prohibitive computational overhead and degrade inference efficiency. Prior studies typically relies on [CLS] attention or text-vision cross-attention to identify and discard redundant visual tokens. Despite promising results, such solutions are prone to introduce positional bias and, more critically, are incompatible with efficient attention kernels such as FlashAttention, limiting their practical deployment for VLM acceleration. In this paper, we step away from attention dependencies and revisit visual token compression from an information-theoretic perspective, aiming to maximally preserve visual information without any attention involvement. We present ApET, an Approximation-Error guided Token compression framework. ApET first reconstructs the original visual tokens with a small set of basis tokens via linear approximation, then leverages the approximation error to identify and drop the least informative tokens. Extensive experiments across multiple VLMs and benchmarks demonstrate that ApET retains 95.2% of the original performance on image-understanding tasks and even attains 100.4% on video-understanding tasks, while compressing the token budgets by 88.9% and 87.5%, respectively. Thanks to its attention-free design, ApET seamlessly integrates with FlashAttention, enabling further inference acceleration and making VLM deployment more practical. Code is available at https://github.com/MaQianKun0/ApET.

Qiankun Ma, Ziyao Zhang, Haofei Wang, Jie Chen, Zhen Song, Hairong Zheng• 2026

Related benchmarks

TaskDatasetResultRank
Visual Question AnsweringVQA v2
Accuracy76.2
1165
Visual Question AnsweringVizWiz
Accuracy51.9
1043
Object Hallucination EvaluationPOPE
Accuracy87.2
935
Multimodal EvaluationMME
Score2.18e+3
557
Visual Question AnsweringGQA
Accuracy63
374
Science Question AnsweringScienceQA (SQA)
Accuracy74.9
128
Vision UnderstandingMMBench
Accuracy63.4
104
Visual Question AnsweringTextVQA
Accuracy57.9
69
Multimodal EvaluationMMBench CN
Accuracy59.3
57
Visual Question AnsweringVQA text
Accuracy54.4
48
Showing 10 of 18 rows

Other info

Follow for update