Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

How Johnny Can Persuade LLMs to Jailbreak Them: Rethinking Persuasion to Challenge AI Safety by Humanizing LLMs

About

Most traditional AI safety research has approached AI models as machines and centered on algorithm-focused attacks developed by security experts. As large language models (LLMs) become increasingly common and competent, non-expert users can also impose risks during daily interactions. This paper introduces a new perspective to jailbreak LLMs as human-like communicators, to explore this overlooked intersection between everyday language interaction and AI safety. Specifically, we study how to persuade LLMs to jailbreak them. First, we propose a persuasion taxonomy derived from decades of social science research. Then, we apply the taxonomy to automatically generate interpretable persuasive adversarial prompts (PAP) to jailbreak LLMs. Results show that persuasion significantly increases the jailbreak performance across all risk categories: PAP consistently achieves an attack success rate of over $92\%$ on Llama 2-7b Chat, GPT-3.5, and GPT-4 in $10$ trials, surpassing recent algorithm-focused attacks. On the defense side, we explore various mechanisms against PAP and, found a significant gap in existing defenses, and advocate for more fundamental mitigation for highly interactive LLMs

Yi Zeng, Hongpeng Lin, Jingwen Zhang, Diyi Yang, Ruoxi Jia, Weiyan Shi• 2024

Related benchmarks

TaskDatasetResultRank
Jailbreak AttackHarmBench
Attack Success Rate (ASR)45.3
376
Jailbreak attack success rateHarmful prompts dataset
Attack Success Rate88
49
JailbreakAdvBench Ensemble configuration GPT-4o
Attack Success Rate (ASR)42
25
Jailbreak AttackAdvBench GPT-3.5-turbo 1.0 (test)
Attack Success Rate87.8
22
Jailbreak AttackAdvbench Llama2-70B Guard 100 prompts Original
ASR50
21
Jailbreak AttackAdvbench Vicuna-33B Guard 100 prompts Original
ASR33
21
Jailbreak AttackJailbreak Evaluation GPT-4o-mini
ASR86.9
13
JailbreakingAdvBench (test)
ASR (GPT-3.5)36
12
Jailbreak AttackClaude 3.5
ASR2
10
Jailbreak AttackAdvbench No Guard 100 prompts Original
ASR8.80e+3
9
Showing 10 of 20 rows

Other info

Follow for update