Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

MoNE: Replacing Redundant Experts with Lightweight Novices for Structured Pruning of MoE

About

Mixture-of-Experts (MoE) enables efficient scaling of large language models by activating only a subset of experts per input token. However, deploying MoE-based models incurs significant memory overhead due to the need to retain all experts in memory. While structured pruning is promising to reduce memory costs, existing methods often show suboptimal performance and unstable degradation in three dimensions: model architectures, calibration data sources, and calibration sample sizes. This paper proposes Mixture-of-Novices-and-Experts (MoNE), a novel expert pruning method that replaces redundant experts with lightweight novices to achieve effective and robust model compression. MoNE evaluates expert redundancy based on two metrics: access frequency and output variance. Experts exhibiting low usage and stable outputs are pruned and replaced with lightweight novices-unbiased estimations of their original outputs-minimizing performance degradation. Extensive experiments demonstrate that MoNE consistently outperforms baseline methods with minimal accuracy degradation across the three dimensions, confirming its effectiveness and robustness. Notably, it outperforms baselines by up to 2.72 for the average zero shot accuracy across nine downstream tasks under 25% pruning ratio, with only 0.14 performance drop for Qwen2-57B-A14B. The code is available at https://github.com/zxgx/mode-pd.

Geng Zhang, Yuxuan Han, Yuxuan Lou, Yiqi Zhang, Wangbo Zhao, Yang You• 2025

Related benchmarks

TaskDatasetResultRank
Question AnsweringARC Challenge
Accuracy56.14
749
Question AnsweringOpenBookQA
Accuracy46.8
465
Question AnsweringARC Easy
Accuracy80.6
386
Natural Language InferenceRTE
Accuracy77.98
367
Boolean Question AnsweringBoolQ
Accuracy85.41
307
Question AnsweringBoolQ
Accuracy89.11
240
Commonsense ReasoningWinoGrande
Accuracy70.17
231
Multitask Language UnderstandingMMLU
Accuracy74.04
206
Common Sense ReasoningCOPA
Accuracy94
138
Question AnsweringOpenBookQA
Accuracy42
84
Showing 10 of 16 rows

Other info

Follow for update