Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

M-RAG: Reinforcing Large Language Model Performance through Retrieval-Augmented Generation with Multiple Partitions

About

Retrieval-Augmented Generation (RAG) enhances Large Language Models (LLMs) by retrieving relevant memories from an external database. However, existing RAG methods typically organize all memories in a whole database, potentially limiting focus on crucial memories and introducing noise. In this paper, we introduce a multiple partition paradigm for RAG (called M-RAG), where each database partition serves as a basic unit for RAG execution. Based on this paradigm, we propose a novel framework that leverages LLMs with Multi-Agent Reinforcement Learning to optimize different language generation tasks explicitly. Through comprehensive experiments conducted on seven datasets, spanning three language generation tasks and involving three distinct language model architectures, we confirm that M-RAG consistently outperforms various baseline methods, achieving improvements of 11%, 8%, and 12% for text summarization, machine translation, and dialogue generation, respectively.

Zheng Wang, Shu Xian Teo, Jieer Ouyang, Yongjun Xu, Wei Shi• 2024

Related benchmarks

TaskDatasetResultRank
SummarizationXsum
ROUGE-224.66
108
SummarizationbigPatent
ROUGE-171.34
61
Machine TranslationJRC-Acquis En-De (dev)
BLEU53.76
18
Machine TranslationJRC-Acquis En-De (test)
BLEU50.75
18
Machine TranslationJRC-Acquis Es-En (dev)
BLEU39.11
18
Machine TranslationJRC-Acquis En-Es (dev)
BLEU35.18
18
Machine TranslationJRC-Acquis De-En (dev)
BLEU49.16
18
SummarizationSummarization dataset
ROUGE-L F130.49
16
Machine TranslationJRC-Acquis Es-En (test)
BLEU39.98
7
Machine TranslationJRC-Acquis En-Es (test)
BLEU32.7
7
Showing 10 of 13 rows

Other info

Follow for update