Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

GAM: Hierarchical Graph-based Agentic Memory for LLM Agents

About

To sustain coherent long-term interactions, Large Language Model (LLM) agents must navigate the tension between acquiring new information and retaining prior knowledge. Current unified stream-based memory systems facilitate context updates but remain vulnerable to interference from transient noise. Conversely, discrete structured memory architectures provide robust knowledge retention but often struggle to adapt to evolving narratives. To address this, we propose GAM, a hierarchical Graph-based Agentic Memory framework that explicitly decouples memory encoding from consolidation to effectively resolve the conflict between rapid context perception and stable knowledge retention. By isolating ongoing dialogue in an event progression graph and integrating it into a topic associative network only upon semantic shifts, our approach minimizes interference while preserving long-term consistency. Additionally, we introduce a graph-guided, multi-factor retrieval strategy to enhance context precision. Experiments on LoCoMo and LongDialQA indicate that our method consistently outperforms state-of-the-art baselines in both reasoning accuracy and efficiency.

Zhaofen Wu, Hanrong Zhang, Fulin Lin, Wujiang Xu, Xinran Xu, Yankai Chen, Henry Peng Zou, Shaowen Chen, Weizhi Zhang, Xue Liu, Philip S. Yu, Hongwei Wang• 2026

Related benchmarks

TaskDatasetResultRank
Long-context Question AnsweringLocomo--
109
Long-context ReasoningLocomo
Average F143.14
45
Temporal ReasoningLocomo
F1 Score51.96
45
Open DomainLocomo
F1 Score28.12
35
Multi-hop ReasoningLocomo
F1 Score35.88
28
Multi-party Dialogue Question AnsweringLongDialQA
F1 Score12.55
28
Overall Reasoning (Average)Locomo
F1 Score (LoCoMo)43.14
28
Single-Hop ReasoningLocomo
F1 Score57.55
28
Proxy multimodal evaluationMovieChat-1K
Accuracy55.51
2
Showing 9 of 9 rows

Other info

Follow for update