Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

CASP: Compression of Large Multimodal Models Based on Attention Sparsity

About

In this work, we propose an extreme compression technique for Large Multimodal Models (LMMs). While previous studies have explored quantization as an efficient post-training compression method for Large Language Models (LLMs), low-bit compression for multimodal models remains under-explored. The redundant nature of inputs in multimodal models results in a highly sparse attention matrix. We theoretically and experimentally demonstrate that the attention matrix's sparsity bounds the compression error of the Query and Key weight matrices. Based on this, we introduce CASP, a model compression technique for LMMs. Our approach performs a data-aware low-rank decomposition on the Query and Key weight matrix, followed by quantization across all layers based on an optimal bit allocation process. CASP is compatible with any quantization technique and enhances state-of-the-art 2-bit quantization methods (AQLM and QuIP#) by an average of 21% on image- and video-language benchmarks.

Mohsen Gholami, Mohammad Akbari, Kevin Cannons, Yong Zhang• 2025

Related benchmarks

TaskDatasetResultRank
Language ModelingWikiText2
Perplexity8.1
1875
Language ModelingC4
Perplexity10.54
1182
Video UnderstandingVideoMME--
192
Multimodal UnderstandingMME--
158
Image CaptioningFlickr30K
CIDEr Score77.2
111
Image CaptioningNoCaps
CIDEr102.1
101
Vision-Language ModelingLiveB
PPL5.69
28
Vision-Language ModelingLWilder
Perplexity4.51
28
Image CaptioningCOCO 17
CIDEr107
23
Image-Language UnderstandingSQA
EM71.2
21
Showing 10 of 18 rows

Other info

Code

Follow for update