Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

LAMP: Extracting Text from Gradients with Language Model Priors

About

Recent work shows that sensitive user data can be reconstructed from gradient updates, breaking the key privacy promise of federated learning. While success was demonstrated primarily on image data, these methods do not directly transfer to other domains such as text. In this work, we propose LAMP, a novel attack tailored to textual data, that successfully reconstructs original text from gradients. Our attack is based on two key insights: (i) modeling prior text probability with an auxiliary language model, guiding the search towards more natural text, and (ii) alternating continuous and discrete optimization, which minimizes reconstruction loss on embeddings, while avoiding local minima by applying discrete text transformations. Our experiments demonstrate that LAMP is significantly more effective than prior work: it reconstructs 5x more bigrams and 23% longer subsequences on average. Moreover, we are the first to recover inputs from batch sizes larger than 1 for textual models. These findings indicate that gradient updates of models operating on textual data leak more information than previously thought.

Mislav Balunovi\'c, Dimitar I. Dimitrov, Nikola Jovanovi\'c, Martin Vechev• 2022

Related benchmarks

TaskDatasetResultRank
Text reconstruction from gradientsRotten Tomatoes
ROUGE-177.6
36
Training Data ReconstructionSST
ROUGE-10.888
32
Training Data ReconstructionRT
ROUGE-10.647
32
Training Data ReconstructionCOLA
ROUGE-189.6
32
Text reconstruction from gradientsCOLA
ROUGE-194.5
24
Text reconstruction from gradientsSST-2
ROUGE-191.6
24
Showing 6 of 6 rows

Other info

Code

Follow for update