CoRect: Context-Aware Logit Contrast for Hidden State Rectification to Resolve Knowledge Conflicts
About
Retrieval-Augmented Generation (RAG) often struggles with knowledge conflicts, where model-internal parametric knowledge overrides retrieved evidence, leading to unfaithful outputs. Existing approaches are often limited, relying either on superficial decoding adjustments or weight editing that necessitates ground-truth targets. Through layer-wise analysis, we attribute this failure to a parametric suppression phenomenon: specifically, in deep layers, certain FFN layers overwrite context-sensitive representations with memorized priors. To address this, we propose CoRect (Context-Aware Logit Contrast for Hidden State Rectification). By contrasting logits from contextualized and non-contextualized forward passes, CoRect identifies layers that exhibit high parametric bias without requiring ground-truth labels. It then rectifies the hidden states to preserve evidence-grounded information. Across question answering (QA) and summarization benchmarks, CoRect consistently improves faithfulness and reduces hallucinations compared to strong baselines.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Abstractive Text Summarization | CNN/Daily Mail (test) | ROUGE-L21.97 | 169 | |
| Question Answering | TriviaQA | EM83 | 116 | |
| Question Answering | SQuAD | Exact Match88.93 | 50 | |
| Abstractive Summarization | XSum (test) | ROUGE-L20.04 | 44 | |
| Question Answering | NQ | EM72.74 | 20 | |
| Question Answering | NQ-Swap | Exact Match80.15 | 20 | |
| Question Answering | HotpotQA | Exact Match45.67 | 20 | |
| Question Answering | TabMWP | EM70.6 | 20 | |
| Abstractive Summarization | TofuEval | Overall Score69.45 | 5 |