Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

APT: Adaptive Pruning and Tuning Pretrained Language Models for Efficient Training and Inference

About

Fine-tuning and inference with large Language Models (LM) are generally known to be expensive. Parameter-efficient fine-tuning over pretrained LMs reduces training memory by updating a small number of LM parameters but does not improve inference efficiency. Structured pruning improves LM inference efficiency by removing consistent parameter blocks, yet often increases training memory and time. To improve both training and inference efficiency, we introduce APT that adaptively prunes and tunes parameters for the LMs. At the early stage of fine-tuning, APT dynamically adds salient tuning parameters for fast and accurate convergence while discarding unimportant parameters for efficiency. Compared to baselines, our experiments show that APT maintains up to 98% task performance when pruning RoBERTa and T5 models with 40% parameters left while keeping 86.4% LLaMA models' performance with 70% parameters remained. Furthermore, APT speeds up LMs fine-tuning by up to 8x and reduces large LMs memory training footprint by up to 70%.

Bowen Zhao, Hannaneh Hajishirzi, Qingqing Cao• 2024

Related benchmarks

TaskDatasetResultRank
Commonsense ReasoningHellaSwag
Accuracy92.6
1460
Commonsense ReasoningWinoGrande
Accuracy81.5
776
Physical Interaction Question AnsweringPIQA
Accuracy85.9
323
Medical Question AnsweringMedMCQA
Accuracy60.7
253
Question AnsweringARC
Accuracy89.1
154
Question AnsweringPubMedQA
Accuracy56.1
145
SummarizationBillSum
Accuracy64.5
28
Financial NLPFinGPT
Accuracy81.3
28
Efficiency EvaluationModel Efficiency Benchmarking Llama3.1-8B
Training Time158
11
Showing 9 of 9 rows

Other info

Follow for update