Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

MA-RAG: Multi-Agent Retrieval-Augmented Generation via Collaborative Chain-of-Thought Reasoning

About

We present MA-RAG, a Multi-Agent framework for Retrieval-Augmented Generation (RAG) that addresses the inherent ambiguities and reasoning challenges in complex information-seeking tasks. Unlike conventional RAG methods that rely on end-to-end fine-tuning or isolated component enhancements, MA-RAG orchestrates a collaborative set of specialized AI agents: Planner, Step Definer, Extractor, and QA Agents, each responsible for a distinct stage of the RAG pipeline. By decomposing tasks into subtasks such as query disambiguation, evidence extraction, and answer synthesis, and enabling agents to communicate intermediate reasoning via chain-of-thought prompting, MA-RAG progressively refines retrieval and synthesis while maintaining modular interpretability. Extensive experiments on multi-hop and ambiguous QA benchmarks, including NQ, HotpotQA, 2WikimQA, and TriviaQA, demonstrate that MA-RAG significantly outperforms standalone LLMs and existing RAG methods across all model scales. Notably, even a small LLaMA3-8B model equipped with MA-RAG surpasses larger standalone LLMs, while larger variants (LLaMA3-70B and GPT-4o-mini) set new state-of-the-art results on challenging multi-hop datasets. Ablation studies reveal that both the planner and extractor agents are critical for multi-hop reasoning, and that high-capacity models are especially important for the QA agent to synthesize answers effectively. Beyond general-domain QA, MA-RAG generalizes to specialized domains such as medical QA, achieving competitive performance against domain-specific models without any domain-specific fine-tuning. Our results highlight the effectiveness of collaborative, modular reasoning in retrieval-augmented systems: MA-RAG not only improves answer accuracy and robustness but also provides interpretable intermediate reasoning steps, establishing a new paradigm for efficient and reliable multi-agent RAG.

Thang Nguyen, Peter Chin, Yu-Wing Tai• 2025

Related benchmarks

TaskDatasetResultRank
Question Answering2WikiMultihopQA
LLM-Acc54.7
20
Question AnsweringMuSiQue
LLM Accuracy40
20
Question AnsweringHotpotQA
LLM Accuracy67.1
20
Long-form Question AnsweringGraphRAG-Bench Med
LLM Accuracy68.3
20
Long-form Question AnsweringNovel GraphRAG-Bench
LLM-Acc45.1
20
Multimodal Document QAVisDoMBench SlideVQA (full)
Accuracy29.4
11
Multimodal Document QAVisDoMBench SPIQA (full)
Accuracy45.52
11
Multimodal Document QAVisDoMBench SciGraphQA (full)
Accuracy29.32
11
Multimodal Document QAVisDoMBench FetaTab (full)
Accuracy27.7
11
Multimodal Document QAVisDoMBench PaperTab (full)
Accuracy33.43
11
Showing 10 of 15 rows

Other info

Follow for update