Talk to Right Specialists: Iterative Routing in Multi-agent Systems for Question Answering
About
Retrieval-augmented generation (RAG) agents are increasingly deployed to answer questions over local knowledge bases that cannot be centralized due to knowledge-sovereignty constraints. This results in two recurring failures in production: users do not know which agent to consult, and complex questions require evidence distributed across multiple agents. To overcome these challenges, we propose RIRS, a training-free orchestration framework to enable a multi-agent system for question answering. In detail, RIRS summarizes each agent's local corpus in an embedding space, enabling a user-facing server to route queries only to the most relevant agents, reducing latency and avoiding noisy "broadcast-to-all" contexts. For complicated questions, the server can iteratively aggregate responses to derive intermediate results and refine the question to bridge the gap toward a comprehensive answer. Extensive experiments demonstrate the effectiveness of RIRS, including its ability to precisely select agents and provide accurate responses to single-hop queries, and its use of an iterative strategy to achieve accurate, multi-step resolutions for complex queries.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Multi-hop Question Answering | 2WikiMultihopQA | EM58.8 | 387 | |
| Multi-hop Question Answering | HotpotQA | F1 Score55.7 | 294 | |
| Multi-hop Question Answering | Multi-hop RAG | F153.7 | 77 | |
| Multi-hop Question Answering | MuSiQue | -- | 24 | |
| Multi-hop Question Answering | 2WikiMultihopQA | Accuracy65.78 | 9 | |
| Multi-hop Question Answering | HotpotQA | Accuracy75.63 | 9 | |
| Question Answering | Natural Questions | LM Score62.05 | 9 | |
| Question Answering | HotpotQA | LM Score54.01 | 9 | |
| Question Answering | 2WikiMultihopQA | LM Score61.14 | 9 | |
| Question Answering | MuSiQue | LM Score24.5 | 9 |