MELoRA: Mini-Ensemble Low-Rank Adapters for Parameter-Efficient Fine-Tuning
About
Parameter-efficient fine-tuning (PEFT) is a popular method for tailoring pre-trained large language models (LLMs), especially as the models' scale and the diversity of tasks increase. Low-rank adaptation (LoRA) is based on the idea that the adaptation process is intrinsically low-dimensional, i.e., significant model changes can be represented with relatively few parameters. However, decreasing the rank encounters challenges with generalization errors for specific tasks when compared to full-parameter fine-tuning. We present MELoRA, a mini-ensemble low-rank adapters that uses fewer trainable parameters while maintaining a higher rank, thereby offering improved performance potential. The core idea is to freeze original pretrained weights and train a group of mini LoRAs with only a small number of parameters. This can capture a significant degree of diversity among mini LoRAs, thus promoting better generalization ability. We conduct a theoretical analysis and empirical studies on various NLP tasks. Our experimental results show that, compared to LoRA, MELoRA achieves better performance with 8 times fewer trainable parameters on natural language understanding tasks and 36 times fewer trainable parameters on instruction following tasks, which demonstrates the effectiveness of MELoRA.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Mathematical Reasoning | SVAMP | Accuracy70 | 368 | |
| Commonsense Reasoning | Common Sense Reasoning Tasks | Avg Score84.72 | 241 | |
| Mathematical Reasoning | MAWPS | Accuracy63.85 | 219 | |
| Mathematical Reasoning | GSM8K | Accuracy48.9 | 212 | |
| Code Generation | HumanEval+ | Pass@117.41 | 189 | |
| Mathematical Reasoning | MATH 500 | Accuracy11.8 | 155 | |
| Mathematical Reasoning | AQUA | Accuracy32.28 | 132 | |
| Natural Language Understanding | GLUE (test val) | MRPC Accuracy90.93 | 59 | |
| Mathematical Reasoning | GSM8K | Accuracy69.36 | 57 | |
| Mathematical Reasoning | MATH500 | Accuracy16.4 | 45 |