Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Resting Neurons, Active Insights: Improving Input Sparsification for Large Language Models

About

Large Language Models (LLMs) achieve state-of-the-art performance across a wide range of applications, but their massive scale poses significant challenges for both efficiency and interpretability. Structural pruning, which reduces model size by removing redundant computational units such as neurons, has been widely explored as a solution, and this study devotes to input sparsification, an increasingly popular technique that improves efficiency by selectively activating only a subset of entry values for each input. However, existing approaches focus primarily on computational savings, often overlooking the representational consequences of sparsification and leaving a noticeable performance gap compared to full models. In this work, we first reinterpret input sparsification as a form of dynamic structural pruning. Motivated by the spontaneous baseline firing rates observed in biological neurons, we introduce a small set of trainable spontaneous neurons that act as compensatory units to stabilize activations in sparsified LLMs. Experiments demonstrate that these auxiliary neurons substantially reduce the sparsification-induced performance gap while generalizing effectively across tasks.

Haotian Xu, Tian Gao, Tsui-Wei Weng, Tengfei Ma• 2025

Related benchmarks

TaskDatasetResultRank
Medical Question AnsweringMedMCQA
Accuracy53.41
253
Question AnsweringCommonsenseQA
Accuracy74.28
143
Question AnsweringTruthfulQA
Accuracy57.15
73
Language ModelingWikitext (test)
Perplexity5.58
52
Question AnsweringMMLU
Accuracy68.79
21
Question AnsweringMathQA
Accuracy46.7
12
Showing 6 of 6 rows

Other info

Follow for update