Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

ASAM: Adaptive Sharpness-Aware Minimization for Scale-Invariant Learning of Deep Neural Networks

About

Recently, learning algorithms motivated from sharpness of loss surface as an effective measure of generalization gap have shown state-of-the-art performances. Nevertheless, sharpness defined in a rigid region with a fixed radius, has a drawback in sensitivity to parameter re-scaling which leaves the loss unaffected, leading to weakening of the connection between sharpness and generalization gap. In this paper, we introduce the concept of adaptive sharpness which is scale-invariant and propose the corresponding generalization bound. We suggest a novel learning method, adaptive sharpness-aware minimization (ASAM), utilizing the proposed generalization bound. Experimental results in various benchmark datasets show that ASAM contributes to significant improvement of model generalization performance.

Jungmin Kwon, Jeongseop Kim, Hyunseo Park, In Kwon Choi• 2021

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-100 (test)
Accuracy89.9
3518
Image ClassificationCIFAR-10 (test)
Accuracy98.68
3381
Image ClassificationCIFAR10 (test)
Accuracy97.56
585
Image ClassificationImageNet (test)--
235
Machine TranslationIWSLT De-En 2014 (test)
BLEU35.02
146
Image ClassificationCIFAR100 (test)
Accuracy56.19
112
Image ClassificationCIFAR10 centralized performance (test)
Accuracy86.07
104
Image ClassificationImageNet Robustness Suite
Top-1 Accuracy (ImageNet-A)92.99
42
Image ClassificationImageNet Clean 1K (val)--
24
Semantic segmentationCityscapes and IDDA (test)
mIoU (Country Seen)46.57
15
Showing 10 of 17 rows

Other info

Follow for update