Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Identify Critical KV Cache in LLM Inference from an Output Perturbation Perspective

About

Large language models have revolutionized natural language processing but face significant challenges of high storage and runtime costs, due to the transformer architecture's reliance on self-attention, particularly the large Key-Value (KV) cache for long-sequence inference. Recent efforts to reduce KV cache size by pruning less critical entries based on attention weights remain empirical and lack formal grounding. This paper presents a formal study on identifying critical KV cache entries by analyzing attention output perturbation. Our analysis reveals that, beyond attention weights, the value states within KV entries and pretrained parameter matrices are also crucial. Based on this, we propose a perturbation-constrained selection algorithm that optimizes the worst-case output perturbation to identify critical entries. Evaluations on the Needle-in-a-Haystack test and Longbench benchmark show our algorithm enhances state-of-the-art cache eviction methods. Further empirical analysis confirms that our algorithm achieves lower output perturbations in over 92% attention heads in Llama model, thereby providing a significant improvement over existing methods.

Yuan Feng, Junlin Lv, Yukun Cao, Xike Xie, S Kevin Zhou• 2025

Related benchmarks

TaskDatasetResultRank
Long-context retrievalRULER 4K-32K context
Accuracy78.9
8
Showing 1 of 1 rows

Other info

Follow for update