Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Improving LoRA in Privacy-preserving Federated Learning

About

Low-rank adaptation (LoRA) is one of the most popular task-specific parameter-efficient fine-tuning (PEFT) methods on pre-trained language models for its good performance and computational efficiency. LoRA injects a product of two trainable rank decomposition matrices over the top of each frozen pre-trained model module. However, when applied in the setting of privacy-preserving federated learning (FL), LoRA may become unstable due to the following facts: 1) the effects of data heterogeneity and multi-step local updates are non-negligible, 2) additive noise enforced on updating gradients to guarantee differential privacy (DP) can be amplified and 3) the final performance is susceptible to hyper-parameters. A key factor leading to these phenomena is the discordance between jointly optimizing the two low-rank matrices by local clients and separately aggregating them by the central server. Thus, this paper proposes an efficient and effective version of LoRA, Federated Freeze A LoRA (FFA-LoRA), to alleviate these challenges and further halve the communication cost of federated fine-tuning LLMs. The core idea of FFA-LoRA is to fix the randomly initialized non-zero matrices and only fine-tune the zero-initialized matrices. Compared to LoRA, FFA-LoRA is motivated by practical and theoretical benefits in privacy-preserved FL. Our experiments demonstrate that FFA-LoRA provides more consistent performance with better computational efficiency over vanilla LoRA in various FL tasks.

Youbang Sun, Zitao Li, Yaliang Li, Bolin Ding• 2024

Related benchmarks

TaskDatasetResultRank
Image ClassificationTiny ImageNet (test)
Accuracy44.62
265
Natural Language UnderstandingGLUE (val)
SST-295.64
170
Question AnsweringSQuAD (test)
F191.07
111
Question AnsweringSQuAD v1.1
F190.31
79
Paraphrase DetectionQQP (test)
Accuracy88.51
51
Commonsense ReasoningCOPA (test)
Accuracy89
46
Natural Language UnderstandingGLUE (test)
SST-2 Accuracy94.32
33
Image ClassificationCIFAR-100 (test)
Accuracy62.1
24
Commonsense ReasoningCommonsense Reasoning Suite (test)
Avg Accuracy0.688
22
Natural Language InferenceGLUE (test)
MNLI Acc91.05
18
Showing 10 of 24 rows

Other info

Follow for update