Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Merging LoRAs like Playing LEGO: Pushing the Modularity of LoRA to Extremes Through Rank-Wise Clustering

About

Low-Rank Adaptation (LoRA) has emerged as a popular technique for fine-tuning large language models (LLMs) to various domains due to its modular design and widespread availability on platforms like Huggingface. This modularity has sparked interest in combining multiple LoRAs to enhance LLM capabilities. However, existing methods for LoRA composition primarily focus on task-specific adaptations that require additional training, and current model merging techniques often fail to fully leverage LoRA's modular nature, leading to parameter interference and performance degradation. In this paper, we investigate the feasibility of disassembling and reassembling multiple LoRAs at a finer granularity, analogous to assembling LEGO blocks. We introduce the concept of Minimal Semantic Units (MSUs), where the parameters corresponding to each rank in LoRA function as independent units. These MSUs demonstrate permutation invariance and concatenation-summation equivalence properties, enabling flexible combinations to create new LoRAs. Building on these insights, we propose the LoRA-LEGO framework. This framework conducts rank-wise parameter clustering by grouping MSUs from different LoRAs into $k$ clusters. The centroid of each cluster serves as a representative MSU, enabling the assembly of a merged LoRA with an adjusted rank of $k$. Additionally, we apply a dual reweighting strategy to optimize the scale of the merged LoRA. Experiments across various benchmarks demonstrate that our method outperforms existing approaches in LoRA merging.

Ziyu Zhao, Tao Shen, Didi Zhu, Zexi Li, Jing Su, Xuwu Wang, Kun Kuang, Fei Wu• 2024

Related benchmarks

TaskDatasetResultRank
Visual Question AnsweringChartQA--
371
Image CaptioningCOCO--
130
Natural Language InferenceQNLI--
61
Visual Question AnsweringIconQA
Top-1 Acc53.1
57
Continual LearningStandard CL Benchmark
Avg Final Acc0.684
50
Continual LearningLarge Number of Tasks
Average Performance56.9
50
Visual Question AnsweringDocVQA (val)
ANLS82.5
47
Image ClassificationDTD
Average Accuracy62.9
19
Image ClassificationEuroSAT
Average Accuracy62.9
19
Multi-task Dense PredictionPascal Context (test)
Normal Error55.4
18
Showing 10 of 29 rows

Other info

Follow for update