Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

AdmTree: Compressing Lengthy Context with Adaptive Semantic Trees

About

The quadratic complexity of self-attention constrains Large Language Models (LLMs) in processing long contexts, a capability essential for many advanced applications. Context compression aims to alleviate this computational bottleneck while retaining critical semantic information. However, existing approaches often fall short: explicit methods may compromise local detail, whereas implicit methods can suffer from positional biases, information degradation, or an inability to capture long-range semantic dependencies. We propose AdmTree, a novel framework for adaptive, hierarchical context compression with a central focus on preserving high semantic fidelity while maintaining efficiency. AdmTree dynamically segments input based on information density, utilizing gist tokens to summarize variable-length segments as the leaves of a semantic binary tree. This structure, together with a lightweight aggregation mechanism and a frozen backbone LLM (thereby minimizing new trainable parameters), enables efficient hierarchical abstraction of the context. By preserving fine-grained details alongside global semantic coherence, mitigating positional bias, and dynamically adapting to content, AdmTree robustly retains the semantic information of long contexts.

Yangning Li, Shaoshen Chen, Yinghui Li, Yankai Chen, Hai-Tao Zheng, Hui Wang, Wenhao Jiang, Philip S. Yu• 2025

Related benchmarks

TaskDatasetResultRank
Mathematical ReasoningGSM8K
Accuracy48.12
983
Long-context Language UnderstandingLongBench (test)
Average Score49.7
133
Long-context UnderstandingLongBench (test)
SingleDoc Performance41.6
30
ReasoningBig-Bench Hard (BBH)
Exact Match (EM)53.53
20
SummarizationArXiv March23
ROUGE-136.1
13
Long-context language modelingLongBench v2
Single Doc QA34.9
10
Multi-turn dialogueShareGPT 1 Turn, 765 tokens
Perplexity4.01
6
Multi-turn dialogueShareGPT 2 Turn, 3006 tokens
PPL2.91
6
Multi-turn dialogueShareGPT 3 Turn 6491 tokens
PPL2.79
6
Showing 9 of 9 rows

Other info

Follow for update