Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Dual LoRA: Enhancing LoRA with Magnitude and Direction Updates

About

Low-rank adaptation (LoRA) is one of the most popular methods among parameter-efficient fine-tuning (PEFT) methods to adapt pre-trained large language models (LLMs) to specific downstream tasks. However, the model trained based on LoRA often has an unsatisfactory performance due to its low-rank assumption. In this paper, we propose a novel method called Dual LoRA to improve the performance by incorporating an inductive bias into the original LoRA. Specifically, we separate low-rank matrices into two groups: the magnitude group to control whether or not and how far we should update a parameter and the direction group to decide whether this parameter should move forward or backward, to better simulate the parameter updating process of the full fine-tuning based on gradient-based optimization algorithms. We show that this can be simply achieved by adding a ReLU function to the magnitude group and a sign function to the direction group. We conduct several experiments over a wide range of NLP tasks, including natural language understanding (NLU) and commonsense reasoning datasets on RoBERTa, DeBERTa, and LLaMA-1/2/3 as baseline models. The results show that we consistently outperform LoRA and its state-of-the-art variants with the same number of trainable parameters.

Yixing Xu, Chao Li, Xuanwu Yin, Spandan Tiwari, Dong Li, Ashish Sirasao, Emad Barsoum• 2025

Related benchmarks

TaskDatasetResultRank
Commonsense ReasoningCommon Sense Reasoning Tasks
Avg Score93
241
Natural language generationE2E NLG Challenge
BLEU70.6
58
Natural Language UnderstandingGLUE
SST-2 Acc97.1
28
Showing 3 of 3 rows

Other info

Follow for update