Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

LoRA-DA: Data-Aware Initialization for Low-Rank Adaptation via Asymptotic Analysis

About

LoRA has become a widely adopted method for PEFT, and its initialization methods have attracted increasing attention. However, existing methods have notable limitations: many methods do not incorporate target-domain data, while gradient-based methods exploit data only at a shallow level by relying on one-step gradient decomposition. In this paper, we establish a theoretical framework for data-aware LoRA initialization. Starting from minimizing the expectation of the parameter discrepancy between the fine-tuned and target models, we derive an optimization problem with two components: a bias term, which is related to the parameter distance between the fine-tuned and target models, and is approximated using a Fisher-gradient formulation to preserve anisotropy; and a variance term, which accounts for the uncertainty introduced by sampling stochasticity through the Fisher information. Solving this problem yields an optimal initialization strategy for LoRA, based on which we develop an efficient algorithm, LoRA-DA. Empirical results across multiple benchmarks demonstrate that LoRA-DA consistently improves final accuracy over existing initialization methods. Additional studies show faster, more stable convergence, robustness across ranks, and only a small initialization overhead for LoRA-DA. The source code will be released upon publication.

Qingyue Zhang, Chang Chu, Tianren Peng, Qi Li, Xiangyang Luo, Zhihao Jiang, Shao-Lun Huang• 2025

Related benchmarks

TaskDatasetResultRank
Math ReasoningMATH
Accuracy12.8
121
Math ReasoningGSM8K
Accuracy (GSM8K)66.4
49
Commonsense ReasoningCommonsense Reasoning Benchmark
BoolQ Accuracy73.2
22
Showing 3 of 3 rows

Other info

Follow for update