Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Can David Beat Goliath? On Multi-Hop Reasoning with Resource-Constrained Agents

About

While reinforcement learning (RL) has empowered multi-turn reasoning agents with retrieval and tools, existing successes largely depend on extensive on-policy rollouts in high-cost, high-accuracy regimes. Under realistic resource constraints that cannot support large models or dense explorations, however, small language model agents fall into a low-cost, low-accuracy regime, where limited rollout budgets lead to sparse exploration, sparse credit assignment, and unstable training. In this work, we challenge this trade-off and show that small language models can achieve strong multi-hop reasoning under resource constraints. We introduce DAVID-GRPO, a budget-efficient RL framework that (i) stabilizes early learning with minimal supervision, (ii) assigns retrieval credit based on evidence recall, and (iii) improves exploration by resampling truncated near-miss trajectories. Evaluated on agents up to 1.5B parameters trained on only four RTX 3090 GPUs, DAVID-GRPO consistently outperforms prior RL methods designed for large-scale settings on six multi-hop QA benchmarks. These results show that with the right inductive biases, small agents can achieve low training cost with high accuracy.

Hojae Han, Heeyun Jung, Jongyoon Kim, Seung-won Hwang• 2026

Related benchmarks

TaskDatasetResultRank
Multi-hop Question AnsweringHotpotQA (test)
F133.8
198
Multi-hop Question Answering2WikiMultiHopQA (test)
EM27.2
143
Multi-hop Question AnsweringMuSiQue (test)
F112.6
111
Multi-hop Question AnsweringBamboogle (test)
EM22
46
Multi-hop Question AnsweringAntileak-m (test)
EM36.3
12
Showing 5 of 5 rows

Other info

Follow for update