Guaranteeing Knowledge Integration with Joint Decoding for Retrieval-Augmented Generation
About
Retrieval-Augmented Generation (RAG) significantly enhances Large Language Models (LLMs) by providing access to external knowledge. However, current research primarily focuses on retrieval quality, often overlooking the critical ''integration bottleneck'': even when relevant documents are retrieved, LLMs frequently fail to utilize them effectively due to conflicts with their internal parametric knowledge. In this paper, we argue that implicitly resolving this conflict in a single generation pass is suboptimal. We introduce GuarantRAG, a framework that explicitly decouples reasoning from evidence integration. First, we generate an ''Inner-Answer'' based solely on parametric knowledge to capture the model's reasoning flow. Second, to guarantee faithful evidence extraction, we generate a ''Refer-Answer'' using a novel Contrastive DPO objective. This objective treats the parametric Inner-Answer as a negative constraint and the retrieved documents as positive ground truth, forcing the model to suppress internal hallucinations in favor of external evidence during this phase. Finally, rather than naive concatenation or using the DPO trained model directly, we propose a joint decoding mechanism that dynamically fuses the logical coherence of the Inner-Answer with the factual precision of the Refer-Answer at the token level. Experiments on five QA benchmarks demonstrate that GuarantRAG improves accuracy by up to 12.1% and reduces hallucinations by 16.3% compared to standard and dynamic RAG baselines.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Question Answering | TruthfulQA | Performance Score81.1 | 52 | |
| Question Answering | Average of 5 datasets | Average Score78.9 | 46 | |
| Retrieval-Augmented Generation | All Datasets Aggregated | Average Performance Score76.6 | 40 | |
| Knowledge Integration Quality | NQ, TruthfulQA, WoW, HotpotQA, ELI5 Aggregate | Average Performance76.7 | 32 | |
| Multi-hop Question Answering | HotpotQA | Performance Score78.2 | 32 | |
| Knowledge Grounded Dialogue | Wizard of Wikipedia (WoW) | Performance Score77.8 | 32 | |
| Long-form Question Answering | ELI5 | Performance Score78.3 | 32 | |
| Question Answering | NQ | Average Performance Score85.7 | 20 | |
| Question Answering | WoW | Average Performance Score84.3 | 20 | |
| Question Answering | HotpotQA | Average Performance Score78.4 | 20 |