Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

SlimGPT: Layer-wise Structured Pruning for Large Language Models

About

Large language models (LLMs) have garnered significant attention for their remarkable capabilities across various domains, whose vast parameter scales present challenges for practical deployment. Structured pruning is an effective method to balance model performance with efficiency, but performance restoration under computational resource constraints is a principal challenge in pruning LLMs. Therefore, we present a low-cost and fast structured pruning method for LLMs named SlimGPT based on the Optimal Brain Surgeon framework. We propose Batched Greedy Pruning for rapid and near-optimal pruning, which enhances the accuracy of head-wise pruning error estimation through grouped Cholesky decomposition and improves the pruning efficiency of FFN via Dynamic Group Size, thereby achieving approximate local optimal pruning results within one hour. Besides, we explore the limitations of layer-wise pruning from the perspective of error accumulation and propose Incremental Pruning Ratio, a non-uniform pruning strategy to reduce performance degradation. Experimental results on the LLaMA benchmark show that SlimGPT outperforms other methods and achieves state-of-the-art results.

Gui Ling, Ziyang Wang, Yuliang Yan, Qingwen Liu• 2024

Related benchmarks

TaskDatasetResultRank
Language ModelingWikiText2
Perplexity7.69
2839
Language ModelingWikiText-2 (test)
PPL17.73
1949
Commonsense ReasoningHellaSwag
Accuracy47.81
1891
Language ModelingWikiText-2
Perplexity (PPL)16.68
1624
Language ModelingC4
Perplexity11.41
1071
Language ModelingPTB
Perplexity37.8
1034
Question AnsweringARC Challenge--
906
Multi-task Language UnderstandingMMLU--
876
Question AnsweringARC Easy
Accuracy58.29
597
Question AnsweringPIQA
Accuracy71.33
374
Showing 10 of 28 rows

Other info

Follow for update