Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Read As Human: Compressing Context via Parallelizable Close Reading and Skimming

About

Large Language Models (LLMs) demonstrate exceptional capability across diverse tasks. However, their deployment in long-context scenarios is hindered by two challenges: computational inefficiency and redundant information. We propose RAM (Read As HuMan), a context compression framework that adopts an adaptive hybrid reading strategy, to address these challenges. Inspired by human reading behavior (i.e., close reading important content while skimming less relevant content), RAM partitions the context into segments and encodes them with the input query in parallel. High-relevance segments are fully retained (close reading), while low-relevance ones are query-guided compressed into compact summary vectors (skimming). Both explicit textual segments and implicit summary vectors are concatenated and fed into decoder to achieve both superior performance and natural language format interpretability. To refine the decision boundary between close reading and skimming, we further introduce a contrastive learning objective based on positive and negative query-segment pairs. Experiments demonstrate that RAM outperforms existing baselines on multiple question answering and summarization benchmarks across two backbones, while delivering up to a 12x end-to-end speedup on long inputs (average length 16K; maximum length 32K).

Jiwei Tang, Shilei Liu, Zhicheng Zhang, Qingsong Lv, Runsong Zhao, Tingwei Lu, Langming Liu, Haibin Chen, Yujin Yuan, Hai-Tao Zheng, Wenbo Su, Bo Zheng• 2026

Related benchmarks

TaskDatasetResultRank
Multi-hop Question Answering2WikiMQA
F1 Score56.47
154
Question AnsweringHotpotQA
EM46.37
79
Question AnsweringNaturalQuestions
EM66.59
39
Question AnsweringNarrativeQA
F132.47
36
SummarizationMultiNews
F1 Score32.07
28
Showing 5 of 5 rows

Other info

Follow for update