Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Byzantine-Robust Distributed Learning: Towards Optimal Statistical Rates

About

In large-scale distributed learning, security issues have become increasingly important. Particularly in a decentralized environment, some computing units may behave abnormally, or even exhibit Byzantine failures -- arbitrary and potentially adversarial behavior. In this paper, we develop distributed learning algorithms that are provably robust against such failures, with a focus on achieving optimal statistical performance. A main result of this work is a sharp analysis of two robust distributed gradient descent algorithms based on median and trimmed mean operations, respectively. We prove statistical error rates for three kinds of population loss functions: strongly convex, non-strongly convex, and smooth non-convex. In particular, these algorithms are shown to achieve order-optimal statistical error rates for strongly convex losses. To achieve better communication efficiency, we further propose a median-based distributed algorithm that is provably robust, and uses only one communication round. For strongly convex quadratic loss, we show that this algorithm achieves the same optimal error rate as the robust distributed gradient descent algorithms.

Dong Yin, Yudong Chen, Kannan Ramchandran, Peter Bartlett• 2018

Related benchmarks

TaskDatasetResultRank
Image ClassificationFashion MNIST (test)
Accuracy84.028
568
Image ClassificationCIFAR-10
Accuracy76
507
Image ClassificationMNIST
Accuracy98
395
Image ClassificationFashion MNIST
Accuracy86
225
Image ClassificationFashionMNIST (test)
Accuracy74.9
218
Image ClassificationMNIST (test)
Test Accuracy92.481
126
Federated Learning Image ClassificationMSTAR
Accuracy95.05
99
Federated Learning Image ClassificationOpenSARShip
ACC91.61
99
Backdoor Attack RobustnessMSTAR (test)
ASR71.12
90
Backdoor Attack RobustnessOpenSARShip (test)
ASR40.81
90
Showing 10 of 50 rows

Other info

Follow for update