Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Injecting Universal Jailbreak Backdoors into LLMs in Minutes

About

Jailbreak backdoor attacks on LLMs have garnered attention for their effectiveness and stealth. However, existing methods rely on the crafting of poisoned datasets and the time-consuming process of fine-tuning. In this work, we propose JailbreakEdit, a novel jailbreak backdoor injection method that exploits model editing techniques to inject a universal jailbreak backdoor into safety-aligned LLMs with minimal intervention in minutes. JailbreakEdit integrates a multi-node target estimation to estimate the jailbreak space, thus creating shortcuts from the backdoor to this estimated jailbreak space that induce jailbreak actions. Our attack effectively shifts the models' attention by attaching strong semantics to the backdoor, enabling it to bypass internal safety mechanisms. Experimental results show that JailbreakEdit achieves a high jailbreak success rate on jailbreak prompts while preserving generation quality, and safe performance on normal queries. Our findings underscore the effectiveness, stealthiness, and explainability of JailbreakEdit, emphasizing the need for more advanced defense mechanisms in LLMs.

Zhuowei Chen, Qiannan Zhang, Shichao Pei• 2025

Related benchmarks

TaskDatasetResultRank
Instruction FollowingAlpacaEval
Win Rate10.5
227
Backdoor AttackDAN (Do-Anything-Now)
ASRw78.79
48
Backdoor AttackMisuse
ASRw68.83
48
Backdoor AttackDNA
ASRw80.23
30
Backdoor AttackDNA (Do-Not-Answer)
ASR (w/ Trigger)66.21
18
Factual AnsweringTruthfulQA
Truthfulness Score56.6
18
Backdoor Attack EvaluationStrongREJECT
ASR (w/ trigger)0.385
18
Mathematical ReasoningGSM-8K
GSM Accuracy72.7
18
Showing 8 of 8 rows

Other info

Follow for update