Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

HydraLoRA: An Asymmetric LoRA Architecture for Efficient Fine-Tuning

About

Adapting Large Language Models (LLMs) to new tasks through fine-tuning has been made more efficient by the introduction of Parameter-Efficient Fine-Tuning (PEFT) techniques, such as LoRA. However, these methods often underperform compared to full fine-tuning, particularly in scenarios involving complex datasets. This issue becomes even more pronounced in complex domains, highlighting the need for improved PEFT approaches that can achieve better performance. Through a series of experiments, we have uncovered two critical insights that shed light on the training and parameter inefficiency of LoRA. Building on these insights, we have developed HydraLoRA, a LoRA framework with an asymmetric structure that eliminates the need for domain expertise. Our experiments demonstrate that HydraLoRA outperforms other PEFT approaches, even those that rely on domain knowledge during the training and inference phases.

Chunlin Tian, Zhan Shi, Zhijiang Guo, Li Li, Chengzhong Xu• 2024

Related benchmarks

TaskDatasetResultRank
Mathematical ReasoningGSM8K (test)
Accuracy62.8
797
Mathematical ReasoningSVAMP
Accuracy71.33
368
Mathematical ReasoningMAWPS
Accuracy65.77
219
Mathematical ReasoningGSM8K
Accuracy64.52
212
Mathematical ReasoningMATH 500
Accuracy14.8
155
Safety EvaluationHEX-PHI
HEx-PHI Score97
148
Mathematical ReasoningAQUA
Accuracy28.35
132
MathGSM8K
Accuracy0.601
87
Commonsense ReasoningOBQA
Accuracy89.2
75
Mathematical ReasoningMATH500
Accuracy16.8
45
Showing 10 of 28 rows

Other info

Follow for update