Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

GIFT-SW: Gaussian noise Injected Fine-Tuning of Salient Weights for LLMs

About

Parameter Efficient Fine-Tuning (PEFT) methods have gained popularity and democratized the usage of Large Language Models (LLMs). Recent studies have shown that a small subset of weights significantly impacts performance. Based on this observation, we introduce a novel PEFT method, called Gaussian noise Injected Fine Tuning of Salient Weights (GIFT-SW). Our method updates only salient columns, while injecting Gaussian noise into non-salient ones. To identify these columns, we developeda generalized sensitivity metric that extends and unifies metrics from previous studies. Experiments with LLaMA models demonstrate that GIFT-SW outperforms full fine-tuning and modern PEFT methods under the same computational budget. Moreover, GIFT-SW offers practical advantages to recover performance of models subjected to mixed-precision quantization with keeping salient weights in full precision.

Maxim Zhelnin, Viktor Moskvoretskii, Egor Shvetsov, Egor Venediktov, Mariya Krylova, Aleksandr Zuev, Evgeny Burnaev• 2024

Related benchmarks

TaskDatasetResultRank
Zero-shot performance evaluationLM Eval Harness (HellaSwag, BoolQ, WinoGrande, PiQA, ARC-easy, ARC-challenge) zero-shot
Mean Accuracy75.46
60
Zero-shot performance evaluationHellaSwag, BoolQ, WinoGrande, PiQA, ARC-easy, and ARC-challenge Zero-shot LM Eval Harness
Mean Accuracy76.78
24
Showing 2 of 2 rows

Other info

Follow for update