Personalized Federated Learning via Variational Bayesian Inference
About
Federated learning faces huge challenges from model overfitting due to the lack of data and statistical diversity among clients. To address these challenges, this paper proposes a novel personalized federated learning method via Bayesian variational inference named pFedBayes. To alleviate the overfitting, weight uncertainty is introduced to neural networks for clients and the server. To achieve personalization, each client updates its local distribution parameters by balancing its construction error over private data and its KL divergence with global distribution from the server. Theoretical analysis gives an upper bound of averaged generalization error and illustrates that the convergence rate of the generalization error is minimax optimal up to a logarithmic factor. Experiments show that the proposed method outperforms other advanced personalized methods on personalized models, e.g., pFedBayes respectively outperforms other SOTA algorithms by 1.25%, 0.42% and 11.71% on MNIST, FMNIST and CIFAR-10 under non-i.i.d. limited data.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | CIFAR10 (test) | Accuracy71.4 | 585 | |
| Classification | fMNIST (test) | Accuracy91.9 | 152 | |
| Image Classification | SUN397 (test) | Top-1 Accuracy83.7 | 136 | |
| Image Classification | EMNIST | Accuracy88.12 | 82 | |
| Personalization | CIFAR-100 (test) | Accuracy78.08 | 80 | |
| Classification | FMNIST | Accuracy81.2 | 53 | |
| Global Prediction | CIFAR-100 original uncorrupted (train) | Accuracy38.39 | 48 | |
| Image Classification | CIFAR-C-100 10 training corruption types (test) | Accuracy82.8 | 48 | |
| Personalization Performance | CIFAR-C-100 (9 held-out corruption types) | Accuracy72.44 | 48 | |
| Global Prediction | CIFAR-100 initial (test) | Accuracy37.23 | 24 |