Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

FLoRA: Federated Fine-Tuning Large Language Models with Heterogeneous Low-Rank Adaptations

About

The rapid development of Large Language Models (LLMs) has been pivotal in advancing AI, with pre-trained LLMs being adaptable to diverse downstream tasks through fine-tuning. Federated learning (FL) further enhances fine-tuning in a privacy-aware manner by utilizing clients' local data through in-situ computation, eliminating the need for data movement. However, fine-tuning LLMs, given their massive scale of parameters, poses challenges for clients with constrained and heterogeneous resources in FL. Previous methods employed low-rank adaptation (LoRA) for efficient federated fine-tuning but utilized traditional FL aggregation strategies on LoRA adapters. These approaches led to mathematically inaccurate aggregation noise, reducing fine-tuning effectiveness and failing to address heterogeneous LoRAs. In this work, we first highlight the mathematical incorrectness of LoRA aggregation in existing federated fine-tuning methods. We introduce a new approach called FLORA that enables federated fine-tuning on heterogeneous LoRA adapters across clients through a novel stacking-based aggregation method. Our approach is noise-free and seamlessly supports heterogeneous LoRA adapters. Extensive experiments demonstrate FLORA' s superior performance in both homogeneous and heterogeneous settings, surpassing state-of-the-art methods. We envision this work as a milestone for efficient, privacy-preserving, and accurate federated fine-tuning of LLMs. Our code is available at https://github.com/ATP-1010/FederatedLLM.

Ziyao Wang, Zheyu Shen, Yexiao He, Guoheng Sun, Hongyi Wang, Lingjuan Lyu, Ang Li• 2024

Related benchmarks

TaskDatasetResultRank
Multi-hop Question Answering2WikiMultihopQA--
278
Multi-hop Question AnsweringHotpotQA--
221
Question AnsweringPopQA--
80
Commonsense ReasoningCommonsense Reasoning (BoolQ, PIQA, SIQA, HellaS., WinoG., ARC-e, ARC-c, OBQA)
BoolQ Accuracy59.45
61
Question AnsweringComplexWebQuestions (CWQ)
Total F134.97
11
Language UnderstandingMMLU (test)
Dolly MMLU Score30.99
10
Multi-turn Conversation EvaluationMT-Bench
Wizard Score4.21
10
Multi-hop Question Answering2WikiMultiHopQA subsampled to 300 Q&A instances
Bridge F147.01
6
Instruction FollowingWizard
MT-Bench Score4.22
5
Instruction FollowingShareGPT
MT-Bench Score3.96
5
Showing 10 of 10 rows

Other info

Code

Follow for update