GenKnowSub: Improving Modularity and Reusability of LLMs through General Knowledge Subtraction
About
Large language models often struggle with zero-shot generalization, and several modular approaches have been proposed to address this challenge. Yet, we hypothesize that a key limitation remains: the entanglement of general knowledge and task-specific adaptations. To overcome this, we propose a modular framework that disentangles these components by constructing a library of task-specific LoRA modules alongside a general-domain LoRA. By subtracting this general knowledge component from each task-specific module, we obtain residual modules that focus more exclusively on task-relevant information, a method we call general knowledge subtraction (GenKnowSub). Leveraging the refined task-specific modules and the Arrow routing algorithm \citep{ostapenko2024towards}, we dynamically select and combine modules for new inputs without additional training. Our studies on the Phi-3 model and standard Arrow as baselines reveal that using general knowledge LoRAs derived from diverse languages, including English, French, and German, yields consistent performance gains in both monolingual and cross-lingual settings across a wide set of benchmarks. Further experiments on Phi-2 demonstrate how GenKnowSub generalizes to weaker LLMs. The complete code and data are available at https://github.com/saharsamr/Modular-LLM.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Reasoning | BBH | Accuracy54 | 507 | |
| Reading Comprehension | BoolQ | Accuracy80.12 | 219 | |
| Reasoning | ARC Easy | Accuracy82.28 | 183 | |
| Reasoning | HellaSwag (HS) | HellaSwag Accuracy74.02 | 142 | |
| Science Question Answering | ARC-E | Accuracy84.38 | 138 | |
| Reasoning | PIQA | Accuracy80.47 | 133 | |
| Science Question Answering | ARC-C | Accuracy56.19 | 127 | |
| Reasoning | WinoGrande (WG) | Accuracy64.72 | 87 | |
| Reasoning | ARC | Accuracy57.19 | 83 | |
| Reasoning | OpenBookQA | Accuracy49.8 | 63 |