Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Neuro-RIT: Neuron-Guided Instruction Tuning for Robust Retrieval-Augmented Language Model

About

Retrieval-Augmented Language Models (RALMs) have demonstrated significant potential in knowledge-intensive tasks; however, they remain vulnerable to performance degradation when presented with irrelevant or noisy retrieved contexts. Existing approaches to enhance robustness typically operate via coarse-grained parameter updates at the layer or module level, often overlooking the inherent neuron-level sparsity of Large Language Models (LLMs). To address this limitation, we propose Neuro-RIT (Neuron-guided Robust Instruction Tuning), a novel framework that shifts the paradigm from dense adaptation to precision-driven neuron alignment. Our method explicitly disentangles neurons that are responsible for processing relevant versus irrelevant contexts using attribution-based neuron mining. Subsequently, we introduce a two-stage instruction tuning strategy that enforces a dual capability for noise robustness: achieving direct noise suppression by functionally deactivating neurons exclusive to irrelevant contexts, while simultaneously optimizing targeted layers for evidence distillation. Extensive experiments across diverse QA benchmarks demonstrate that Neuro-RIT consistently outperforms strong baselines and robustness-enhancing methods.

Jaemin Kim, Jae O Lee, Sumyeong Ahn, Seo Yeon Park• 2026

Related benchmarks

TaskDatasetResultRank
Question Answering2Wiki--
152
Question AnsweringPopQA
Accuracy74.29
52
Question AnsweringASQA
Accuracy80.48
51
Question AnsweringNQ
Exact Match72.57
46
Question AnsweringHotpotQA
Accuracy55.71
37
Showing 5 of 5 rows

Other info

Follow for update