Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Prompt Injection attack against LLM-integrated Applications

About

Large Language Models (LLMs), renowned for their superior proficiency in language comprehension and generation, stimulate a vibrant ecosystem of applications around them. However, their extensive assimilation into various services introduces significant security risks. This study deconstructs the complexities and implications of prompt injection attacks on actual LLM-integrated applications. Initially, we conduct an exploratory analysis on ten commercial applications, highlighting the constraints of current attack strategies in practice. Prompted by these limitations, we subsequently formulate HouYi, a novel black-box prompt injection attack technique, which draws inspiration from traditional web injection attacks. HouYi is compartmentalized into three crucial elements: a seamlessly-incorporated pre-constructed prompt, an injection prompt inducing context partition, and a malicious payload designed to fulfill the attack objectives. Leveraging HouYi, we unveil previously unknown and severe attack outcomes, such as unrestricted arbitrary LLM usage and uncomplicated application prompt theft. We deploy HouYi on 36 actual LLM-integrated applications and discern 31 applications susceptible to prompt injection. 10 vendors have validated our discoveries, including Notion, which has the potential to impact millions of users. Our investigation illuminates both the possible risks of prompt injection attacks and the possible tactics for mitigation.

Yi Liu, Gelei Deng, Yuekang Li, Kailong Wang, Zihao Wang, Xiaofeng Wang, Tianwei Zhang, Yepang Liu, Haoyu Wang, Yan Zheng, Leo Yu Zhang, Yang Liu• 2023

Related benchmarks

TaskDatasetResultRank
Open-domain Question AnsweringMS Marco--
48
Adversarial Attack2Wiki
ASR59
30
Fact-Level RAG Poisoning AttackFinQA
RSR@599.8
7
RAG Poisoning Attack (Document-Level Targeting)FinQA
RSR@543.7
7
Fact-Level RAG Poisoning AttackTiEBe
RSR@598.6
7
RAG Poisoning Attack (Document-Level Targeting)BioASQ
RSR@533.8
7
RAG Poisoning Attack (Document-Level Targeting)TiEBe
RSR@592.4
7
Fact-Level RAG Poisoning AttackBioASQ
RSR@559.6
7
Code GenerationCode
F1 Score72
4
Showing 9 of 9 rows

Other info

Follow for update