Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Implicit Regularization of Sharpness-Aware Minimization for Scale-Invariant Problems

About

Sharpness-aware minimization (SAM) improves generalization of various deep learning tasks. Motivated by popular architectures such as LoRA, we explore the implicit regularization of SAM for scale-invariant problems involving two groups of variables. Instead of focusing on commonly used sharpness, this work introduces a concept termed balancedness, defined as the difference between the squared norm of two variables. This allows us to depict richer global behaviors of SAM. In particular, our theoretical and empirical findings reveal that i) SAM promotes balancedness; and ii) the regularization on balancedness is data-responsive -- outliers have stronger impact. The latter coincides with empirical observations that SAM outperforms SGD in the presence of outliers. Leveraging the implicit regularization, we develop a resource-efficient SAM variant, balancedness-aware regularization (BAR), tailored for scale-invariant problems such as finetuning language models with LoRA. BAR saves 95% computational overhead of SAM, with enhanced test performance across various tasks on RoBERTa, GPT2, and OPT-1.3B.

Bingcong Li, Liang Zhang, Niao He• 2024

Related benchmarks

TaskDatasetResultRank
Natural Language InferenceSNLI (test)
Accuracy84.9
681
Natural Language UnderstandingGLUE (test)
SST-2 Accuracy96
416
Sentiment ClassificationSST2 (test)
Accuracy91.5
214
Sentiment AnalysisSST-5 (test)
Accuracy55
173
Question ClassificationTREC (test)
Accuracy96.7
124
Natural Language InferenceRTE (test)
Accuracy81
52
Data-to-text generationWebNLG (test)
BLEU55.2
39
Natural Language InferenceMNLI (test)
Accuracy0.783
38
Few-shot and zero-shot learning evaluation across multiple NLP tasksSST-2, CB, RTE, COPA, ReCoRD, SQuAD (test)
SST-2 Accuracy93.7
6
Natural Language UnderstandingGLUE (test)
STS-B Score92.6
5
Showing 10 of 10 rows

Other info

Code

Follow for update