Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

MTSA: Multi-turn Safety Alignment for LLMs through Multi-round Red-teaming

About

The proliferation of jailbreak attacks against large language models (LLMs) highlights the need for robust security measures. However, in multi-round dialogues, malicious intentions may be hidden in interactions, leading LLMs to be more prone to produce harmful responses. In this paper, we propose the \textbf{M}ulti-\textbf{T}urn \textbf{S}afety \textbf{A}lignment (\ourapproach) framework, to address the challenge of securing LLMs in multi-round interactions. It consists of two stages: In the thought-guided attack learning stage, the red-team model learns about thought-guided multi-round jailbreak attacks to generate adversarial prompts. In the adversarial iterative optimization stage, the red-team model and the target model continuously improve their respective capabilities in interaction. Furthermore, we introduce a multi-turn reinforcement learning algorithm based on future rewards to enhance the robustness of safety alignment. Experimental results show that the red-team model exhibits state-of-the-art attack capabilities, while the target model significantly improves its performance on safety benchmarks.

Weiyang Guo, Jing Li, Wenya Wang, YU LI, Daojing He, Jun Yu, Min Zhang• 2025

Related benchmarks

TaskDatasetResultRank
Jailbreak AttackHarmBench
Attack Success Rate (ASR)84.5
376
Jailbreak AttackJailbreakBench (JBB)
ASR52.73
54
General PerformanceAlpacaEval
Winrate77.45
25
LLM Safety DefenseMTSA-R3
ASR23.5
20
Response Quality EvaluationMT-Bench
Average Response Quality6.78
19
JailbreakingAdvBench (test)
ASR (GPT-3.5)72
12
Adversarial RobustnessRedQueen (out-of-domain)
ASR19.5
4
General LLM EvaluationXSTest
Refusal Rate23.1
4
LLM Safety DefenseBeavertails
ASR10.39
4
LLM Safety DefenseCoSafe
ASR15.42
4
Showing 10 of 11 rows

Other info

Code

Follow for update