Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Self-Correcting RAG: Enhancing Faithfulness via MMKP Context Selection and NLI-Guided MCTS

About

Retrieval-augmented generation (RAG) substantially extends the knowledge boundary of large language models. However, it still faces two major challenges when handling complex reasoning tasks: low context utilization and frequent hallucinations. To address these issues, we propose Self-Correcting RAG, a unified framework that reformulates retrieval and generation as constrained optimization and path planning. On the input side, we move beyond traditional greedy retrieval and, for the first time, formalize context selection as a multi-dimensional multiple-choice knapsack problem (MMKP), thereby maximizing information density and removing redundancy under a strict token budget. On the output side, we introduce a natural language inference (NLI)-guided Monte Carlo Tree Search (MCTS) mechanism, which leverages test-time compute to dynamically explore reasoning trajectories and validate the faithfulness of generated answers. Experiments on six multi-hop question answering and fact-checking datasets demonstrate that our method significantly improves reasoning accuracy on complex queries while effectively reducing hallucinations, outperforming strong existing baselines.Our code is available at https://github.com/xjiacs/Self-Correcting-RAG .

Shijia Xu, Zhou Wu, Xiaolong Jia, Yu Wang, Kai Liu, April Xiaowen Dong• 2026

Related benchmarks

TaskDatasetResultRank
Multi-hop Question Answering2Wiki
Exact Match31.2
152
Multi-hop Question AnsweringMulti-hop RAG
F144.8
77
Question AnsweringNQ
EM48.4
69
RetrievalHotpotQA
R@593.6
36
RetrievalNQ
R@572.8
19
RetrievalPopQA
R@565.5
19
Retrieval2Wiki
Recall@586.9
19
Multi-hop Question AnsweringMuSiQue
Exact Match (EM)22.7
12
Question AnsweringAverage (NQ, PopQA, MuSiQue, 2Wiki, HotpotQA, MultiHop-RAG)
EM37.1
12
Simple Question AnsweringPopQA
EM43.2
12
Showing 10 of 10 rows

Other info

Follow for update