Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

PASER: Post-Training Data Selection for Efficient Pruned Large Language Model Recovery

About

Model pruning is an effective approach for compressing large language models (LLMs). However, this process often leads to significant degradation of model capabilities. While post-training techniques such as instruction tuning are commonly employed to recover model performance, existing methods often overlook the uneven deterioration of model capabilities and incur high computational costs. Moreover, some irrelevant instructions may also introduce negative effects to model capacity recovery. To address these challenges, we propose the \textbf{P}ost-training d\textbf{A}ta \textbf{S}election method for \textbf{E}fficient pruned large language model \textbf{R}ecovery (\textbf{PASER}). PASER aims to identify instructions to recover the most compromised model capacities with a certain data budget. Our approach first applies manifold learning and spectral clustering to group recovery instructions in the semantic space, revealing capability-specific instruction sets. Then, the data budget is adaptively allocated across clusters by the degree of corresponding model capability degradation. In each cluster, we prioritize data samples that lead to the most decline of model performance. To mitigate potential negative tuning effects, we also detect and filter out conflicting or irrelevant recovery data. Extensive experiments demonstrate that PASER significantly outperforms conventional baselines, effectively recovering the general capabilities of pruned LLMs while utilizing merely 4\%-20\% of the original post-training data. We provide the code repository in \href{https://github.com/BokwaiHo/PASER}{Link}.

Bowei He, Lihao Yin, Hui-Ling Zhen, Xiaokun Zhang, Mingxuan Yuan, Chen Ma• 2025

Related benchmarks

TaskDatasetResultRank
Language ModelingWikiText2
Perplexity7.24
1875
Language ModelingWikiText-2 (test)
PPL11.87
1541
Commonsense ReasoningHellaSwag
Accuracy77.05
1460
Language ModelingWikiText-2
Perplexity (PPL)11.25
841
Commonsense ReasoningWinoGrande
Accuracy69.97
776
Language ModelingPTB
Perplexity14.16
650
Commonsense ReasoningPIQA
Accuracy79.84
647
Language ModelingPTB (test)
Perplexity22.63
471
Question AnsweringOBQA
Accuracy42.9
276
Question AnsweringARC-E
Accuracy69.14
242
Showing 10 of 27 rows

Other info

Follow for update