Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Language Model Inversion

About

Language models produce a distribution over the next token; can we use this information to recover the prompt tokens? We consider the problem of language model inversion and show that next-token probabilities contain a surprising amount of information about the preceding text. Often we can recover the text in cases where it is hidden from the user, motivating a method for recovering unknown prompts given only the model's current distribution output. We consider a variety of model access scenarios, and show how even without predictions for every token in the vocabulary we can recover the probability vector through search. On Llama-2 7b, our inversion method reconstructs prompts with a BLEU of $59$ and token-level F1 of $78$ and recovers $27\%$ of prompts exactly. Code for reproducing all experiments is available at http://github.com/jxmorris12/vec2text.

John X. Morris, Wenting Zhao, Justin T. Chiu, Vitaly Shmatikov, Alexander M. Rush• 2023

Related benchmarks

TaskDatasetResultRank
Prompt RecoveryArxiv Math
BLEU-135
14
Prompt RecoveryAlpaca
BLEU-119.71
14
Prompt RecoverySelf-Instruct
BLEU-122.14
14
Showing 3 of 3 rows

Other info

Follow for update