Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Context Compression via Explicit Information Transmission

About

Long-context inference with Large Language Models (LLMs) is costly due to quadratic attention and growing key-value caches, motivating context compression. In this work, we study soft context compression, where a long context is condensed into a small set of continuous representations. Existing methods typically re-purpose the LLM itself as a trainable compressor, relying on layer-by-layer self-attention to iteratively aggregate information. We argue that this paradigm suffers from two structural limitations: (i) progressive representation overwriting across layers (ii) uncoordinated allocation of compression capacity across tokens. We propose ComprExIT (Context Compression via Explicit Information Transmission), a lightweight framework that formulates soft compression into a new paradigm: explicit information transmission over frozen LLM hidden states. This decouples compression from the model's internal self-attention dynamics. ComprExIT performs (i) depth-wise transmission to selectively transmit multi-layer information into token anchors, mitigating progressive overwriting, and (ii) width-wise transmission to aggregate anchors into a small number of slots via a globally optimized transmission plan, ensuring coordinated allocation of information. Across six question-answering benchmarks, ComprExIT consistently outperforms state-of-the-art context compression methods while introducing only ~1% additional parameters, demonstrating that explicit and coordinated information transmission enables more effective and robust long-context compression.

Jiangnan Ye, Hanqi Yan, Zhenyi Shen, Heng Chang, Ye Mao, Yulan He• 2026

Related benchmarks

TaskDatasetResultRank
Question AnsweringSQuAD
F175.68
127
Question AnsweringHotpotQA
F174.15
114
Question AnsweringTriviaQA
F178.86
46
Question AnsweringTextbookQA MRQA out-of-domain evaluation
EM56.09
29
Question AnsweringRelExt MRQA out-of-domain evaluation
EM65.31
29
Question AnsweringSearchQA
EM78
14
Reading ComprehensionDROP MRQA out-of-domain
EM0.4358
14
Question AnsweringNQ
EM52.23
14
Reading ComprehensionBioASQ MRQA out-of-domain
EM59.64
14
Reading ComprehensionDuoRC MRQA out-of-domain
EM34.31
14
Showing 10 of 11 rows

Other info

Follow for update