Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

The Good and The Bad: Exploring Privacy Issues in Retrieval-Augmented Generation (RAG)

About

Retrieval-augmented generation (RAG) is a powerful technique to facilitate language model with proprietary and private data, where data privacy is a pivotal concern. Whereas extensive research has demonstrated the privacy risks of large language models (LLMs), the RAG technique could potentially reshape the inherent behaviors of LLM generation, posing new privacy issues that are currently under-explored. In this work, we conduct extensive empirical studies with novel attack methods, which demonstrate the vulnerability of RAG systems on leaking the private retrieval database. Despite the new risk brought by RAG on the retrieval data, we further reveal that RAG can mitigate the leakage of the LLMs' training data. Overall, we provide new insights in this paper for privacy protection of retrieval-augmented LLMs, which benefit both LLMs and RAG systems builders. Our code is available at https://github.com/phycholosogy/RAG-privacy.

Shenglai Zeng, Jiankun Zhang, Pengfei He, Yue Xing, Yiding Liu, Han Xu, Jie Ren, Shuaiqiang Wang, Dawei Yin, Yi Chang, Jiliang Tang• 2024

Related benchmarks

TaskDatasetResultRank
Subgraph Reconstruction AttackENRON
Precision19.8
56
Subgraph Reconstruction AttackHCM
Precision9.7
56
Data Extraction AttackEHRAgent
Equality (EQ)14
20
Data Extraction AttackReAct
EQ13
20
Data Extraction AttackRAP
Equality (EQ)12
20
Importance-based Node LeakageAgriculture
Leakage (Degree)53.8
10
Graph Extraction AttackM-GraphRAG Medical 1.0 (test)
Leak (Nodes)67.84
10
Importance-based Node Leakagemedical
Leakage (Deg)83.9
10
Graph Extraction AttackAgriculture LightRAG 1.0 (test)
Leakage (N)37.49
5
Knowledge Base ExtractionChatDoctor Pirates attack
CRR95
5
Showing 10 of 18 rows

Other info

Code

Follow for update