Locally Differentially Private Document Generation Using Zero Shot Prompting
About
Numerous studies have highlighted the privacy risks associated with pretrained large language models. In contrast, our research offers a unique perspective by demonstrating that pretrained large language models can effectively contribute to privacy preservation. We propose a locally differentially private mechanism called DP-Prompt, which leverages the power of pretrained large language models and zero-shot prompting to counter author de-anonymization attacks while minimizing the impact on downstream utility. When DP-Prompt is used with a powerful language model like ChatGPT (gpt-3.5), we observe a notable reduction in the success rate of de-anonymization attacks, showing that it surpasses existing approaches by a considerable margin despite its simpler design. For instance, in the case of the IMDB dataset, DP-Prompt (with ChatGPT) perfectly recovers the clean sentiment F1 score while achieving a 46\% reduction in author identification F1 score against static attackers and a 26\% reduction against adaptive attackers. We conduct extensive experiments across six open-source large language models, ranging up to 7 billion parameters, to analyze various effects of the privacy-utility tradeoff.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Multi-hop Question Answering | 2WikiMultihopQA | -- | 387 | |
| Multi-hop Question Answering | HotpotQA | -- | 294 | |
| Commonsense Question Answering | CSQA | Accuracy6.49 | 58 | |
| Question Answering | SQuAD | Score7.2 | 29 | |
| Multi-task Language Understanding | MMLU | MMLU Accuracy59 | 14 | |
| Multi-hop Question Answering | MuSiQue | F1 Score8 | 14 | |
| Reasoning Question Answering | StrategyQA | Accuracy56 | 14 | |
| Clinical Downstream Task | Pri-DDX | Accuracy46.48 | 12 | |
| Clinical Downstream Task | Pri-DDX, Pri-NLICE, and Pri-SLJA Aggregate | Average Accuracy37.57 | 12 | |
| Clinical Downstream Task | Pri-SLJA | Accuracy37.7 | 12 |