Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

MoRE: A Mixture of Low-Rank Experts for Adaptive Multi-Task Learning

About

With the rapid development of Large Language Models (LLMs), Parameter-Efficient Fine-Tuning (PEFT) methods have gained significant attention, which aims to achieve efficient fine-tuning of LLMs with fewer parameters. As a representative PEFT method, Low-Rank Adaptation (LoRA) introduces low-rank matrices to approximate the incremental tuning parameters and achieves impressive performance over multiple scenarios. After that, plenty of improvements have been proposed for further improvement. However, these methods either focus on single-task scenarios or separately train multiple LoRA modules for multi-task scenarios, limiting the efficiency and effectiveness of LoRA in multi-task scenarios. To better adapt to multi-task fine-tuning, in this paper, we propose a novel Mixture of Low-Rank Experts (MoRE) for multi-task PEFT. Specifically, instead of using an individual LoRA for each task, we align different ranks of LoRA module with different tasks, which we named low-rank experts. Moreover, we design a novel adaptive rank selector to select the appropriate expert for each task. By jointly training low-rank experts, MoRE can enhance the adaptability and efficiency of LoRA in multi-task scenarios. Finally, we conduct extensive experiments over multiple multi-task benchmarks along with different LLMs to verify model performance. Experimental results demonstrate that compared to traditional LoRA and its variants, MoRE significantly improves the performance of LLMs in multi-task scenarios and incurs no additional inference cost. We also release the model and code to facilitate the community.

Dacao Zhang, Kun Zhang, Shimao Chu, Le Wu, Xin Li, Si Wei• 2025

Related benchmarks

TaskDatasetResultRank
Natural Language UnderstandingGLUE
SST-293.9
531
Common Sense ReasoningBoolQ
Accuracy74.7
212
Commonsense ReasoningARC-C
Accuracy64.5
172
Commonsense ReasoningOBQA
Accuracy80.5
117
Commonsense ReasoningARC-E
Accuracy80
106
Action UnderstandingMMT-47 Action Understanding
Accuracy53.48
17
High-Level ReasoningMMT-47 High Level Reasoning
Accuracy45.58
17
Commonsense ReasoningMMT-47 Commonsense Reasoning
Accuracy84.44
17
Vision UnderstandingMMT-47 Vision Benchmark
Accuracy77.98
17
Object Motion and Spatial ReasoningMMT-47 Object Motion & Spatial
Accuracy64.95
17
Showing 10 of 12 rows

Other info

Follow for update