Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

THINKSAFE: Self-Generated Safety Alignment for Reasoning Models

About

Large reasoning models (LRMs) achieve remarkable performance by leveraging reinforcement learning (RL) on reasoning tasks to generate long chain-of-thought (CoT) reasoning. However, this over-optimization often prioritizes compliance, making models vulnerable to harmful prompts. To mitigate this safety degradation, recent approaches rely on external teacher distillation, yet this introduces a distributional discrepancy that degrades native reasoning. We propose ThinkSafe, a self-generated alignment framework that restores safety alignment without external teachers. Our key insight is that while compliance suppresses safety mechanisms, models often retain latent knowledge to identify harm. ThinkSafe unlocks this via lightweight refusal steering, guiding the model to generate in-distribution safety reasoning traces. Fine-tuning on these self-generated responses effectively realigns the model while minimizing distribution shift. Experiments on DeepSeek-R1-Distill and Qwen3 show ThinkSafe significantly improves safety while preserving reasoning proficiency. Notably, it achieves superior safety and comparable reasoning to GRPO, with significantly reduced computational cost. Code, models, and datasets are available at https://github.com/seanie12/ThinkSafe.git.

Seanie Lee, Sangwoo Park, Yumin Choi, Gyeongman Kim, Minki Kang, Jihun Yun, Dongmin Park, Jongho Park, Sung Ju Hwang• 2026

Related benchmarks

TaskDatasetResultRank
Mathematical ReasoningMATH 500
pass@191.9
153
Mathematical ReasoningAIME 2024
Pass@151.25
54
Safety EvaluationStrongREJECT
Attack Success Rate26.52
45
Over-refusalXSTest--
42
ReasoningReasoning Evaluation Suite AIME 2024, GSM8k, MATH 500, GPQA
AIME 2024 Score0.7333
32
Safety EvaluationSafety Evaluation Suite HarmBench, StrongReject, WildJailbreak, XSTest
HarmBench Score40.37
28
Harmfulness EvaluationHarmBench
Harmful Response Ratio27.08
21
SafetyWildJailbreak
Harmful Response Ratio21.15
21
ReasoningGSM8K
Pass@190.1
21
ReasoningGPQA
Pass@146.28
21
Showing 10 of 10 rows

Other info

GitHub

Follow for update