Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Neural networks with late-phase weights

About

The largely successful method of training neural networks is to learn their weights using some variant of stochastic gradient descent (SGD). Here, we show that the solutions found by SGD can be further improved by ensembling a subset of the weights in late stages of learning. At the end of learning, we obtain back a single model by taking a spatial average in weight space. To avoid incurring increased computational costs, we investigate a family of low-dimensional late-phase weight models which interact multiplicatively with the remaining parameters. Our results show that augmenting standard models with late-phase weights improves generalization in established benchmarks such as CIFAR-10/100, ImageNet and enwik8. These findings are complemented with a theoretical analysis of a noisy quadratic problem which provides a simplified picture of the late phases of neural network learning.

Johannes von Oswald, Seijin Kobayashi, Alexander Meulemans, Christian Henning, Benjamin F. Grewe, Jo\~ao Sacramento• 2020

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-100 (test)
Accuracy85
3518
Image ClassificationCIFAR-10 (test)
Accuracy97.45
3381
Character-level Language Modelingenwik8 (test)
BPC1.615
195
Out-of-Distribution DetectionCIFAR-10
AUROC83.4
105
Out-of-Distribution DetectionSVHN
AUROC87.7
62
Out-of-Distribution DetectionLSUN
AUROC0.884
26
Character-level Language Modelingenwik8 (train)
BPC1.522
12
Out-of-Distribution DetectionTiny ImageNet (Out-of-distribution) vs CIFAR-100 (In-distribution)
OOD AUROC86.24
10
Out-of-Distribution DetectionTiny ImageNet (TIN)
AUROC0.863
10
Robustness EvaluationCIFAR-100-C
mCE43.15
8
Showing 10 of 10 rows

Other info

Code

Follow for update