Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

AutoDAN: Generating Stealthy Jailbreak Prompts on Aligned Large Language Models

About

The aligned Large Language Models (LLMs) are powerful language understanding and decision-making tools that are created through extensive alignment with human feedback. However, these large models remain susceptible to jailbreak attacks, where adversaries manipulate prompts to elicit malicious outputs that should not be given by aligned LLMs. Investigating jailbreak prompts can lead us to delve into the limitations of LLMs and further guide us to secure them. Unfortunately, existing jailbreak techniques suffer from either (1) scalability issues, where attacks heavily rely on manual crafting of prompts, or (2) stealthiness problems, as attacks depend on token-based algorithms to generate prompts that are often semantically meaningless, making them susceptible to detection through basic perplexity testing. In light of these challenges, we intend to answer this question: Can we develop an approach that can automatically generate stealthy jailbreak prompts? In this paper, we introduce AutoDAN, a novel jailbreak attack against aligned LLMs. AutoDAN can automatically generate stealthy jailbreak prompts by the carefully designed hierarchical genetic algorithm. Extensive evaluations demonstrate that AutoDAN not only automates the process while preserving semantic meaningfulness, but also demonstrates superior attack strength in cross-model transferability, and cross-sample universality compared with the baseline. Moreover, we also compare AutoDAN with perplexity-based defense methods and show that AutoDAN can bypass them effectively.

Xiaogeng Liu, Nan Xu, Muhao Chen, Chaowei Xiao• 2023

Related benchmarks

TaskDatasetResultRank
Jailbreak AttackHarmBench
Attack Success Rate (ASR)80.5
376
Jailbreak AttackAdvBench
AASR86.7
247
Adversarial Attack Success RateAdvBench
ASR24.04
75
JailbreakAdvBench
Avg Queries28.6
63
Jailbreak AttackJailbreakBench
ASR17
54
JailbreakHarmBench Standard Behaviours (200 examples)
ASR0.00e+0
48
Red-teaming Safety EvaluationStrongREJECT
ASR20.45
32
Autonomous DrivingAgent-Driver
Accuracy (ACC)90.7
24
Knowledge-intensive QAStrategyQA
ACC56.1
24
Healthcare Record ManagementEHRAgent
Accuracy (ACC)68.4
24
Showing 10 of 35 rows

Other info

Follow for update