Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Locally Differentially Private Document Generation Using Zero Shot Prompting

About

Numerous studies have highlighted the privacy risks associated with pretrained large language models. In contrast, our research offers a unique perspective by demonstrating that pretrained large language models can effectively contribute to privacy preservation. We propose a locally differentially private mechanism called DP-Prompt, which leverages the power of pretrained large language models and zero-shot prompting to counter author de-anonymization attacks while minimizing the impact on downstream utility. When DP-Prompt is used with a powerful language model like ChatGPT (gpt-3.5), we observe a notable reduction in the success rate of de-anonymization attacks, showing that it surpasses existing approaches by a considerable margin despite its simpler design. For instance, in the case of the IMDB dataset, DP-Prompt (with ChatGPT) perfectly recovers the clean sentiment F1 score while achieving a 46\% reduction in author identification F1 score against static attackers and a 26\% reduction against adaptive attackers. We conduct extensive experiments across six open-source large language models, ranging up to 7 billion parameters, to analyze various effects of the privacy-utility tradeoff.

Saiteja Utpala, Sara Hooker, Pin Yu Chen• 2023

Related benchmarks

TaskDatasetResultRank
Multi-hop Question Answering2WikiMultihopQA--
387
Multi-hop Question AnsweringHotpotQA--
294
Commonsense Question AnsweringCSQA
Accuracy6.49
58
Question AnsweringSQuAD
Score7.2
29
Multi-task Language UnderstandingMMLU
MMLU Accuracy59
14
Multi-hop Question AnsweringMuSiQue
F1 Score8
14
Reasoning Question AnsweringStrategyQA
Accuracy56
14
Clinical Downstream TaskPri-DDX
Accuracy46.48
12
Clinical Downstream TaskPri-DDX, Pri-NLICE, and Pri-SLJA Aggregate
Average Accuracy37.57
12
Clinical Downstream TaskPri-SLJA
Accuracy37.7
12
Showing 10 of 18 rows

Other info

Follow for update