Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Defending against Backdoors in Federated Learning with Robust Learning Rate

About

Federated learning (FL) allows a set of agents to collaboratively train a model without sharing their potentially sensitive data. This makes FL suitable for privacy-preserving applications. At the same time, FL is susceptible to adversarial attacks due to decentralized and unvetted data. One important line of attacks against FL is the backdoor attacks. In a backdoor attack, an adversary tries to embed a backdoor functionality to the model during training that can later be activated to cause a desired misclassification. To prevent backdoor attacks, we propose a lightweight defense that requires minimal change to the FL protocol. At a high level, our defense is based on carefully adjusting the aggregation server's learning rate, per dimension and per round, based on the sign information of agents' updates. We first conjecture the necessary steps to carry a successful backdoor attack in FL setting, and then, explicitly formulate the defense based on our conjecture. Through experiments, we provide empirical evidence that supports our conjecture, and we test our defense against backdoor attacks under different settings. We observe that either backdoor is completely eliminated, or its accuracy is significantly reduced. Overall, our experiments suggest that our defense significantly outperforms some of the recently proposed defenses in the literature. We achieve this by having minimal influence over the accuracy of the trained models. In addition, we also provide convergence rate analysis for our proposed scheme.

Mustafa Safa Ozdayi, Murat Kantarcioglu, Yulia R. Gel• 2020

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-100 (test)
Accuracy44.35
3518
Image ClassificationTiny ImageNet (test)--
362
Backdoor DefenseCIFAR-10 (test)
Clean Accuracy75.26
58
Federated Time Series ForecastingFHWA
MSE0.182
45
Federated Time Series ForecastingPDCCH
MSE0.0978
45
Image ClassificationCIFAR-10 IID
Average BA0.0346
37
Model Poisoning DefensePDCCH
MSE0.0061
36
Model Poisoning DefenseFHWA
MSE0.0254
36
Sentiment AnalysisSentiment140
Mean Accuracy51.23
14
Backdoor DefenseCIFAR-10 non-IID (test)
Clean MA59.87
13
Showing 10 of 25 rows

Other info

Follow for update