Adaptive Federated Optimization
About
Federated learning is a distributed machine learning paradigm in which a large number of clients coordinate with a central server to learn a model without sharing their own training data. Standard federated optimization methods such as Federated Averaging (FedAvg) are often difficult to tune and exhibit unfavorable convergence behavior. In non-federated settings, adaptive optimization methods have had notable success in combating such issues. In this work, we propose federated versions of adaptive optimizers, including Adagrad, Adam, and Yogi, and analyze their convergence in the presence of heterogeneous data for general non-convex settings. Our results highlight the interplay between client heterogeneity and communication efficiency. We also perform extensive experiments on these methods and show that the use of adaptive optimizers can significantly improve the performance of federated learning.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | CIFAR10 (test) | Accuracy70 | 585 | |
| Image Classification | Tiny ImageNet (test) | Accuracy44.81 | 265 | |
| Skin lesion classification | HAM10000 (test) | Accuracy83.22 | 83 | |
| Inference Attack | Federated Learning environments Unauthorized Access (test) | Inference Attack Accuracy8.6 | 66 | |
| Inference Attack | Federated Learning environments Authorized Access (test) | Inference Attack Accuracy30.12 | 66 | |
| Regression | PovertyMap (test) | Worst-U/R Pearson Correlation0.7294 | 43 | |
| Federated Learning | ACSIncome | Local Average Distance (AD)0.23 | 30 | |
| Wildlife Species Classification | WILDS-iWildCam ID (test) | Macro F135.7 | 23 | |
| Image Classification | Cifar10 Dirichlet(0.3) (test) | -- | 21 | |
| Image Classification | CIFAR-100 Dirichlet 0.6, 500 clients, 2% participation (test) | Accuracy (500R)37.57 | 13 |