Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Demystifying Reinforcement Learning in Agentic Reasoning

About

Recently, the emergence of agentic RL has showcased that RL could also effectively improve the agentic reasoning ability of LLMs, yet the key design principles and optimal practices remain unclear. In this work, we conduct a comprehensive and systematic investigation to demystify reinforcement learning in agentic reasoning from three key perspectives: data, algorithm, and reasoning mode. We highlight our key insights: (i) Replacing stitched synthetic trajectories with real end-to-end tool-use trajectories yields a far stronger SFT initialization; high-diversity, model-aware datasets sustain exploration and markedly improve RL performance. (ii) Exploration-friendly techniques are crucial for agentic RL, such as clip higher, overlong reward shaping, and maintaining adequate policy entropy could improve the training efficiency. (iii) A deliberative strategy with fewer tool calls outperforms frequent tool calls or verbose self-reasoning, improving tool efficiency and final accuracy. Together, these simple practices consistently enhance agentic reasoning and training efficiency, achieving strong results on challenging benchmarks with smaller models, and establishing a practical baseline for future agentic RL research. Beyond these empirical insights, we further contribute a high-quality, real end-to-end agentic SFT dataset along with a high-quality RL dataset, and demonstrate the effectiveness of our insights in boosting the agentic reasoning ability of LLMs across four challenging benchmarks, including AIME2024/AIME2025, GPQA-Diamond, and LiveCodeBench-v6. With our recipes, 4B-sized models could also achieve superior agentic reasoning performance compared to 32B-sized models. Code and models: https://github.com/Gen-Verse/Open-AgentRL

Zhaochen Yu, Ling Yang, Jiaru Zou, Shuicheng Yan, Mengdi Wang• 2025

Related benchmarks

TaskDatasetResultRank
Mathematical ReasoningMATH
Accuracy91.6
535
Mathematical ReasoningAIME 25
Accuracy70
201
Scientific Question AnsweringGPQA Diamond
Accuracy58.5
64
Mathematical ReasoningHMMT 2025--
38
CodingLiveCodeBench
Task Accuracy26.8
23
Mathematical ReasoningAIME 24
Accuracy72.6
17
Mathematical ReasoningAIME 2025
Avg Score @1670
14
Mathematical ReasoningAIME 2024
AIME 2024 Avg Score72.6
14
Mathematical ReasoningBeyondAIME
avg@1635.3
8
Showing 9 of 9 rows

Other info

GitHub

Follow for update