Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Making Them Ask and Answer: Jailbreaking Large Language Models in Few Queries via Disguise and Reconstruction

About

In recent years, large language models (LLMs) have demonstrated notable success across various tasks, but the trustworthiness of LLMs is still an open problem. One specific threat is the potential to generate toxic or harmful responses. Attackers can craft adversarial prompts that induce harmful responses from LLMs. In this work, we pioneer a theoretical foundation in LLMs security by identifying bias vulnerabilities within the safety fine-tuning and design a black-box jailbreak method named DRA (Disguise and Reconstruction Attack), which conceals harmful instructions through disguise and prompts the model to reconstruct the original harmful instruction within its completion. We evaluate DRA across various open-source and closed-source models, showcasing state-of-the-art jailbreak success rates and attack efficiency. Notably, DRA boasts a 91.1% attack success rate on OpenAI GPT-4 chatbot.

Tong Liu, Yingjie Zhang, Zhe Zhao, Yinpeng Dong, Guozhu Meng, Kai Chen• 2024

Related benchmarks

TaskDatasetResultRank
Jailbreak AttackMalicious goals dataset (test)
ASR8.25
99
JailbreakJBB-Behaviors utilitarian dilemmas (test)
Jailbreak Success Rate83
72
Jailbreak AttackJailbreakBench (JBB)--
62
Jailbreak AttackAdvBench-50 + Malicious Instruct
ASR98
40
Jailbreak AttackHARMFULQA
JADES31
33
Jailbreak Attack Stealth EvaluationAdvBench 50
PPL14.6255
10
Stealth Adversarial AttackGemini-3
Perplexity17.8293
9
Malicious Code GenerationMalicious Code Generation Dataset (test)
ASR (Claude3.5)0.00e+0
7
Jailbreak AttackGlm-4
ASR0.9417
5
Jailbreak AttackVicuna
ASR90.83
5
Showing 10 of 15 rows

Other info

Follow for update