Meta-Reinforcement Learning with Self-Reflection for Agentic Search
About
This paper introduces MR-Search, an in-context meta reinforcement learning (RL) formulation for agentic search with self-reflection. Instead of optimizing a policy within a single independent episode with sparse rewards, MR-Search trains a policy that conditions on past episodes and adapts its search strategy across episodes. MR-Search learns to learn a search strategy with self-reflection, allowing search agents to improve in-context exploration at test-time. Specifically, MR-Search performs cross-episode exploration by generating explicit self-reflections after each episode and leveraging them as additional context to guide subsequent attempts, thereby promoting more effective exploration during test-time. We further introduce a multi-turn RL algorithm that estimates a dense relative advantage at the turn level, enabling fine-grained credit assignment on each episode. Empirical results across various benchmarks demonstrate the advantages of MR-Search over baselines based RL, showing strong generalization and relative improvements of 9.2% to 19.3% across eight benchmarks. Our code and data are available at https://github.com/tengxiao1/MR-Search.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Multi-hop Question Answering | HotpotQA 2018 Wikipedia dump (dev) | Accuracy46.8 | 14 | |
| Multi-hop Question Answering | 2wiki 2018 Wikipedia dump (dev) | Accuracy (%)43.6 | 14 | |
| Multi-hop Question Answering | Musique 2018 Wikipedia dump (dev) | Accuracy22.1 | 14 | |
| Multi-hop Question Answering | Bamboogle 2018 Wikipedia dump (dev) | Accuracy45.2 | 14 | |
| Single-hop Question Answering | NQ (Natural Questions) 2018 Wikipedia dump (dev) | Accuracy50.2 | 14 | |
| Single-hop Question Answering | TriviaQA 2018 Wikipedia dump (dev) | Accuracy66.6 | 14 | |
| Single-hop Question Answering | PopQA 2018 Wikipedia dump (dev) | Accuracy47.2 | 14 |