Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Abusing Images and Sounds for Indirect Instruction Injection in Multi-Modal LLMs

About

We demonstrate how images and sounds can be used for indirect prompt and instruction injection in multi-modal LLMs. An attacker generates an adversarial perturbation corresponding to the prompt and blends it into an image or audio recording. When the user asks the (unmodified, benign) model about the perturbed image or audio, the perturbation steers the model to output the attacker-chosen text and/or make the subsequent dialog follow the attacker's instruction. We illustrate this attack with several proof-of-concept examples targeting LLaVa and PandaGPT.

Eugene Bagdasaryan, Tsung-Yin Hsieh, Ben Nassi, Vitaly Shmatikov• 2023

Related benchmarks

TaskDatasetResultRank
Visual Question AnsweringVQA v2
Refusal Rate (Exact)65
5
Showing 1 of 1 rows

Other info

Follow for update