Learning-Rate-Free Learning by D-Adaptation
About
D-Adaptation is an approach to automatically setting the learning rate which asymptotically achieves the optimal rate of convergence for minimizing convex Lipschitz functions, with no back-tracking or line searches, and no additional function value or gradient evaluations per step. Our approach is the first hyper-parameter free method for this class without additional multiplicative log factors in the convergence rate. We present extensive experiments for SGD and Adam variants of our method, where the method automatically matches hand-tuned learning rates across more than a dozen diverse machine learning problems, including large-scale vision and language problems. An open-source implementation is available.
Aaron Defazio, Konstantin Mishchenko• 2023
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | CIFAR-10 (test) | Accuracy89.6 | 3381 | |
| Image Classification | ImageNet-100 (test) | Clean Accuracy76.9 | 109 | |
| Image Classification | Food-101 (test) | -- | 89 | |
| Language Modeling | C4 LLaMA-130M (val) | Perplexity18.672 | 27 | |
| Image Classification | CIFAR-10 | Latency (ms/iter)24.16 | 13 | |
| Image Classification | MNIST (test) | Accuracy99.58 | 12 | |
| Instance Segmentation | MS-COCO 2017 (test) | Box mAP5057.4 | 6 | |
| Keypoint Detection | MS-COCO 2017 (test) | mAP50 (Box)80.3 | 6 | |
| Molecular property prediction | OGBG | mAP22.1 | 6 | |
| MRI Reconstruction | fastMRI | SSIM0.722 | 6 |
Showing 10 of 10 rows