AdmTree: Compressing Lengthy Context with Adaptive Semantic Trees
About
The quadratic complexity of self-attention constrains Large Language Models (LLMs) in processing long contexts, a capability essential for many advanced applications. Context compression aims to alleviate this computational bottleneck while retaining critical semantic information. However, existing approaches often fall short: explicit methods may compromise local detail, whereas implicit methods can suffer from positional biases, information degradation, or an inability to capture long-range semantic dependencies. We propose AdmTree, a novel framework for adaptive, hierarchical context compression with a central focus on preserving high semantic fidelity while maintaining efficiency. AdmTree dynamically segments input based on information density, utilizing gist tokens to summarize variable-length segments as the leaves of a semantic binary tree. This structure, together with a lightweight aggregation mechanism and a frozen backbone LLM (thereby minimizing new trainable parameters), enables efficient hierarchical abstraction of the context. By preserving fine-grained details alongside global semantic coherence, mitigating positional bias, and dynamically adapting to content, AdmTree robustly retains the semantic information of long contexts.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Mathematical Reasoning | GSM8K | Accuracy48.12 | 983 | |
| Long-context Language Understanding | LongBench (test) | Average Score49.7 | 133 | |
| Long-context Understanding | LongBench (test) | SingleDoc Performance41.6 | 30 | |
| Reasoning | Big-Bench Hard (BBH) | Exact Match (EM)53.53 | 20 | |
| Summarization | ArXiv March23 | ROUGE-136.1 | 13 | |
| Long-context language modeling | LongBench v2 | Single Doc QA34.9 | 10 | |
| Multi-turn dialogue | ShareGPT 1 Turn, 765 tokens | Perplexity4.01 | 6 | |
| Multi-turn dialogue | ShareGPT 2 Turn, 3006 tokens | PPL2.91 | 6 | |
| Multi-turn dialogue | ShareGPT 3 Turn 6491 tokens | PPL2.79 | 6 |