Dynamic Privacy Allocation for Locally Differentially Private Federated Learning with Composite Objectives
About
This paper proposes a locally differentially private federated learning algorithm for strongly convex but possibly nonsmooth problems that protects the gradients of each worker against an honest but curious server. The proposed algorithm adds artificial noise to the shared information to ensure privacy and dynamically allocates the time-varying noise variance to minimize an upper bound of the optimization error subject to a predefined privacy budget constraint. This allows for an arbitrarily large but finite number of iterations to achieve both privacy protection and utility up to a neighborhood of the optimal solution, removing the need for tuning the number of iterations. Numerical results show the superiority of the proposed algorithm over state-of-the-art methods.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | CIFAR-100 (test) | Accuracy35.11 | 3518 | |
| Image Classification | CIFAR-10 (test) | Accuracy79.36 | 3381 | |
| Image Classification | CIFAR-100 | Accuracy83.57 | 302 | |
| Image Classification | FEMNIST (test) | Accuracy73.18 | 104 | |
| Image Classification | CIFAR-10 v1 (test) | Accuracy79.36 | 98 | |
| Image Classification | CIFAR-10 | Noise Scale9.03e+3 | 48 | |
| Image Classification | CIFAR-100 | Noise Scale3.79e+5 | 48 | |
| Image Classification | CIFAR10 (train) | Training Time (s)8.96e+4 | 41 | |
| Image Classification | CIFAR-100 v1 (test) | Accuracy (%)53.04 | 24 | |
| Differential Privacy Noise Scale Evaluation | CIFAR-10 | Noise Scale (L2 norm)3.86e+5 | 24 |