Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

PRAISE: Prefix-Based Rollout Reuse in Agentic Search Training

About

In agentic search, large language models (LLMs) are trained to perform multi-turn retrieval and reasoning for complex tasks such as multi-hop question answering (QA). However, current search-based Reinforcement Learning (RL) methods suffer from two core limitations: expensive long-horizon rollouts are under-utilized during training, and supervision is typically available only at the final answer, resulting in severe reward sparsity. We present Prefix-based Rollout reuse for Agentic search with Intermediate Step rEwards (PRAISE), a framework for improving both data efficiency and credit assignment in agentic search training. Given a complete search trajectory, PRAISE extracts prefix states at different search turns, elicits intermediate answers from them, and uses these prefixes both to construct additional training trajectories and to derive step-level rewards from performance differences across prefixes. Our method uses a single shared model for both search policy learning and prefix answer evaluation, enabling joint optimization without extra human annotations or a separate reward model. Experiments on multi-hop QA benchmarks show that PRAISE consistently improves performance over strong baselines.

Erhan Zhang, Yiqun Chen, Zechun Niu, Wei Yang, Xiaochi Wei, Yan Gao, Yi Wu, Yao Hu, Jiaxin Mao• 2026

Related benchmarks

TaskDatasetResultRank
Question Answering2Wiki
F158.14
152
Question AnsweringBamboogle
EM48
120
Question AnsweringNQ
F143.62
9
Question AnsweringHotpotQA
F1 Score60.62
9
Question AnsweringMuSiQue
F1 Score30.73
9
Showing 5 of 5 rows

Other info

Follow for update