Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Lookahead Optimizer: k steps forward, 1 step back

About

The vast majority of successful deep neural networks are trained using variants of stochastic gradient descent (SGD) algorithms. Recent attempts to improve SGD can be broadly categorized into two approaches: (1) adaptive learning rate schemes, such as AdaGrad and Adam, and (2) accelerated schemes, such as heavy-ball and Nesterov momentum. In this paper, we propose a new optimization algorithm, Lookahead, that is orthogonal to these previous approaches and iteratively updates two sets of weights. Intuitively, the algorithm chooses a search direction by looking ahead at the sequence of fast weights generated by another optimizer. We show that Lookahead improves the learning stability and lowers the variance of its inner optimizer with negligible computation and memory cost. We empirically demonstrate Lookahead can significantly improve the performance of SGD and Adam, even with their default hyperparameter settings on ImageNet, CIFAR-10/100, neural machine translation, and Penn Treebank.

Michael R. Zhang, James Lucas, Geoffrey Hinton, Jimmy Ba• 2019

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-100 (val)
Accuracy78.34
661
Language ModelingPenn Treebank (test)
Perplexity57.72
411
Image ClassificationSVHN--
359
Image ClassificationImageNet (val)
Top-1 Accuracy75.49
354
Image ClassificationCIFAR-10 (val)
Top-1 Accuracy95.27
329
Image ClassificationCIFAR-100--
302
Image ClassificationGTSRB
Accuracy54.22
291
Image ClassificationImageNet (val)
Top-1 Accuracy76.52
188
Language ModelingPenn Treebank (val)
Perplexity60.28
178
Machine TranslationWMT English-German 2014 (test)
BLEU24.7
136
Showing 10 of 16 rows

Other info

Code

Follow for update