FedCM: Federated Learning with Client-level Momentum
About
Federated Learning is a distributed machine learning approach which enables model training without data sharing. In this paper, we propose a new federated learning algorithm, Federated Averaging with Client-level Momentum (FedCM), to tackle problems of partial participation and client heterogeneity in real-world federated learning applications. FedCM aggregates global gradient information in previous communication rounds and modifies client gradient descent with a momentum-like term, which can effectively correct the bias and improve the stability of local SGD. We provide theoretical analysis to highlight the benefits of FedCM. We also perform extensive empirical studies and demonstrate that FedCM achieves superior performance in various tasks and is robust to different levels of client numbers, participation rate and client heterogeneity.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | CIFAR-100 (test) | Accuracy51.01 | 3518 | |
| Image Classification | TinyImageNet (test) | Accuracy41.37 | 366 | |
| Image Classification | Tiny ImageNet (test) | Accuracy84.24 | 265 | |
| Image Classification | OfficeHome | -- | 131 | |
| Language Modeling | C4 | -- | 73 | |
| Image Classification | CIFAR-100 (test) | Communication Rounds120 | 61 | |
| Image Classification | CIFAR-10 IID | Accuracy84.22 | 58 | |
| Image Classification | TinyImageNet (test) | Communication Rounds342 | 56 | |
| Image Classification | CIFAR-100 IID | Accuracy71.12 | 37 | |
| Image Classification | Tiny-ImageNet Dirichlet-0.05 (test) | Accuracy36 | 32 |