TreeRPO: Tree Relative Policy Optimization
About
Large Language Models (LLMs) have shown remarkable reasoning capabilities through Reinforcement Learning with Verifiable Rewards (RLVR) methods. However, a key limitation of existing approaches is that rewards defined at the full trajectory level provide insufficient guidance for optimizing the intermediate steps of a reasoning process. To address this, we introduce \textbf{\name}, a novel method that estimates the mathematical expectations of rewards at various reasoning steps using tree sampling. Unlike prior methods that rely on a separate step reward model, \name directly estimates these rewards through this sampling process. Building on the group-relative reward training mechanism of GRPO, \name innovatively computes rewards based on step-level groups generated during tree sampling. This advancement allows \name to produce fine-grained and dense reward signals, significantly enhancing the learning process and overall performance of LLMs. Experimental results demonstrate that our \name algorithm substantially improves the average Pass@1 accuracy of Qwen-2.5-Math on test benchmarks, increasing it from 19.0\% to 35.5\%. Furthermore, \name significantly outperforms GRPO by 2.9\% in performance while simultaneously reducing the average response length by 18.1\%, showcasing its effectiveness and efficiency. Our code will be available at \href{https://github.com/yangzhch6/TreeRPO}{https://github.com/yangzhch6/TreeRPO}.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Multi-hop Question Answering | 2WikiMQA | F1 Score70.1 | 154 | |
| Multi-hop Question Answering | MuSiQue | -- | 106 | |
| Single-hop Question Answering | TriviaQA | -- | 62 | |
| Single-hop Question Answering | PopQA | -- | 55 | |
| Multi-hop Question Answering | HotpotQA | F1 Score59.6 | 31 | |
| Multi-hop Question Answering | Bamboogle | F155.7 | 25 | |
| Question Answering | Knowledge-Intensive Question Answering Benchmarks Aggregate | F156.5 | 15 |