Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Compressing LLMs with MoP: Mixture of Pruners

About

The high computational demands of Large Language Models (LLMs) motivate methods that reduce parameter count and accelerate inference. In response, model pruning emerges as an effective strategy, yet current methods typically focus on a single dimension-depth or width. We introduce MoP (Mixture of Pruners), an iterative framework that unifies these dimensions. At each iteration, MoP generates two branches-pruning in depth versus pruning in width-and selects a candidate to advance the path. On LLaMA-2 and LLaMA-3, MoP advances the frontier of structured pruning, exceeding the accuracy of competing methods across a broad set of compression regimes. It also consistently outperforms depth-only and width-only pruning. Furthermore, MoP translates structural pruning into real speedup, reducing end-to-end latency by 39% at 40% compression. Finally, extending MoP to the vision-language model LLaVA-1.5, we notably improve computational efficiency and demonstrate that text-only recovery fine-tuning can restore performance even on visual tasks.

Bruno Lopes Yamamoto, Lucas Lauton de Alcantara, Victor Zacarias, Leandro Giusti Mugnaini, Keith Ando Ogawa, Lucas Pellicer, Rosimeire Pereira Costa, Edson Bollis, Anna Helena Reali Costa, Artur Jordao• 2026

Related benchmarks

TaskDatasetResultRank
Question AnsweringARC Challenge
Accuracy44.62
749
Question AnsweringARC Easy
Accuracy70.54
386
Question AnsweringWinoGrande (WG)
Accuracy66.8
98
Question AnsweringPIQA
Accuracy76.35
83
Multiple-choice Question AnsweringHellaSwag
Accuracy71.94
59
Question AnsweringWinoGrande, HellaSwag, ARC-e, ARC-c, PIQA Average
Avg Accuracy65.44
35
Showing 6 of 6 rows

Other info

Follow for update