Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

ADAM: A Systematic Data Extraction Attack on Agent Memory via Adaptive Querying

About

Large Language Model (LLM) agents have achieved rapid adoption and demonstrated remarkable capabilities across a wide range of applications. To improve reasoning and task execution, modern LLM agents would incorporate memory modules or retrieval-augmented generation (RAG) mechanisms, enabling them to further leverage prior interactions or external knowledge. However, such a design also introduces a group of critical privacy vulnerabilities: sensitive information stored in memory can be leaked through query-based attacks. Although feasible, existing attacks often achieve only limited performance, with low attack success rates (ASR). In this paper, we propose ADAM, a novel privacy attack that features data distribution estimation of a victim agent's memory and employs an entropy-guided query strategy for maximizing privacy leakage. Extensive experiments demonstrate that our attack substantially outperforms state-of-the-art ones, achieving up to 100% ASRs. These results thus underscore the urgent need for robust privacy-preserving methods for current LLM agents.

Xingyu Lyu, Jianfeng He, Ning Wang, Yidan Hu, Tao Li, Danjue Chen, Shixiong Li, Yimin Chen• 2026

Related benchmarks

TaskDatasetResultRank
Data Extraction AttackEHRAgent
Equality (EQ)83
20
Data Extraction AttackReAct
EQ86
20
Data Extraction AttackRAP
Equality (EQ)73
20
Data Extraction Attack on Agent MemoryEhrAgent (test)
Equality (EQ)82
12
Data Extraction Attack on Agent MemoryReAct (test)
EQ Score81
12
Data Extraction Attack on Agent MemoryRAP (test)
Equality Score71
12
Showing 6 of 6 rows

Other info

Follow for update