Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Multi-Layered Memory Architectures for LLM Agents: An Experimental Evaluation of Long-Term Context Retention

About

Long-horizon dialogue systems suffer from semanticdrift and unstable memory retention across extended sessions. This paper presents a Multi-Layer Memory Framework that decomposes dialogue history into working, episodic, and semantic layers with adaptive retrieval gating and retention regularization. The architecture controls cross-session drift while maintaining bounded context growth and computational efficiency. Experiments on LOCOMO, LOCCO, and LoCoMo show improved performance, achieving 46.85 Success Rate, 0.618 overall F1 with 0.594 multi-hop F1, and 56.90% six-period retention while reducing false memory rate to 5.1% and context usage to 58.40%. Results confirm enhanced long-term retention and reasoning stability under constrained context budgets.

Sunil Tiwari, Payal Fofadiya• 2026

Related benchmarks

TaskDatasetResultRank
Long-horizon dialogueLOCCO
SR99.1
5
Long-horizon dialogueLocomo
Success Rate46.85
4
Structured ReasoningLocomo
F1 Score61.8
2
Long-context UnderstandingLongBench--
1
Showing 4 of 4 rows

Other info

Follow for update