Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

R-Search: Empowering LLM Reasoning with Search via Multi-Reward Reinforcement Learning

About

Large language models (LLMs) have notably progressed in multi-step and long-chain reasoning. However, extending their reasoning capabilities to encompass deep interactions with search remains a non-trivial challenge, as models often fail to identify optimal reasoning-search interaction trajectories, resulting in suboptimal responses. We propose R-Search, a novel reinforcement learning framework for Reasoning-Search integration, designed to enable LLMs to autonomously execute multi-step reasoning with deep search interaction, and learn optimal reasoning search interaction trajectories via multi-reward signals, improving response quality in complex logic- and knowledge-intensive tasks. R-Search guides the LLM to dynamically decide when to retrieve or reason, while globally integrating key evidence to enhance deep knowledge interaction between reasoning and search. During RL training, R-Search provides multi-stage, multi-type rewards to jointly optimize the reasoning-search trajectory. Experiments on seven datasets show that R-Search outperforms advanced RAG baselines by up to 32.2% (in-domain) and 25.1% (out-of-domain). The code and data are available at https://github.com/QingFei1/R-Search.

Qingfei Zhao, Ruobing Wang, Dingling Xu, Daren Zha, Limin Liu• 2025

Related benchmarks

TaskDatasetResultRank
Multi-hop Question Answering2WikiMultihopQA
EM32
387
Multi-hop Question AnsweringBamboogle
Exact Match20.8
128
Multi-hop Question AnsweringHotpotQA
Exact Match (EM)30.7
117
Question AnsweringNQ (Natural Questions)
EM31.9
70
Multi-hop Question AnsweringMuSiQue
Exact Match (EM)11.9
58
General Question AnsweringTriviaQA
Exact Match54.1
54
General Question AnsweringPopQA
EM36.5
51
Showing 7 of 7 rows

Other info

Follow for update