Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Towards Safer Large Reasoning Models by Promoting Safety Decision-Making before Chain-of-Thought Generation

About

Large reasoning models (LRMs) achieved remarkable performance via chain-of-thought (CoT), but recent studies showed that such enhanced reasoning capabilities are at the expense of significantly degraded safety capabilities. In this paper, we reveal that LRMs' safety degradation occurs only after CoT is enabled, and this degradation is not observed when CoT is disabled. This observation motivates us to consider encouraging LRMs to make safety decisions before CoT generation. To this end, we propose a novel safety alignment method that promotes the safety decision-making of LRMs before starting CoT generation. Specifically, we first utilize a Bert-based classifier to extract safety decision signals from a safe model (e.g., a CoT-disabled LRM) and then integrate these signals into LRMs' safety alignment as auxiliary supervision. In this way, the safety gradients can be backpropagated to the LRMs' latent representations, effectively strengthening the LRMs' safety decision-making abilities against CoT generation. Extensive experiments demonstrate that our method substantially improves the safety capabilities of LRMs while effectively maintaining LRMs' general reasoning performance.

Jianan Chen, Zhifang Zhang, Shuo He, Linan Yue, Lei Feng, Minling Zhang• 2026

Related benchmarks

TaskDatasetResultRank
ReasoningGPQA Diamond
Accuracy46.3
135
Jailbreak DefenseWild Jailbreak
ASR14
114
Jailbreak DefensePAIR
ASR0.00e+0
97
Jailbreak DefenseGCG
ASR0.00e+0
91
CodingHumanEval+
Pass@160.3
83
Safety EvaluationStrongREJECT
Attack Success Rate5.4
65
Jailbreak DefenseJBC
ASR0.00e+0
54
Jailbreak DefenseStrongREJECT
Attack Success Rate2.9
54
Safety EvaluationWildJailbreak
ASR0.188
53
CodingMBPP+
Pass@148.4
52
Showing 10 of 13 rows

Other info

Follow for update