Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

LoRA Done RITE: Robust Invariant Transformation Equilibration for LoRA Optimization

About

Low-rank adaption (LoRA) is a widely used parameter-efficient finetuning method for LLM that reduces memory requirements. However, current LoRA optimizers lack transformation invariance, meaning the actual updates to the weights depends on how the two LoRA factors are scaled or rotated. This deficiency leads to inefficient learning and sub-optimal solutions in practice. This paper introduces LoRA-RITE, a novel adaptive matrix preconditioning method for LoRA optimization, which can achieve transformation invariance and remain computationally efficient. We provide theoretical analysis to demonstrate the benefit of our method and conduct experiments on various LLM tasks with different models including Gemma 2B, 7B, and mT5-XXL. The results demonstrate consistent improvements against existing optimizers. For example, replacing Adam with LoRA-RITE during LoRA fine-tuning of Gemma-2B yielded 4.6\% accuracy gain on Super-Natural Instructions and 3.5\% accuracy gain across other four LLM benchmarks (HellaSwag, ArcChallenge, GSM8K, OpenBookQA).

Jui-Nan Yen, Si Si, Zhao Meng, Felix Yu, Sai Surya Duvvuri, Inderjit S. Dhillon, Cho-Jui Hsieh, Sanjiv Kumar• 2024

Related benchmarks

TaskDatasetResultRank
Question AnsweringARC-E
Accuracy88.26
416
Question AnsweringOBQA
Accuracy83.4
300
Common Sense ReasoningBoolQ
Accuracy74.19
212
Social Interaction Question AnsweringSIQA
Accuracy80.25
109
Multiple-choice Question AnsweringHellaSwag
Accuracy93.21
93
Question AnsweringARC-C
Accuracy71.5
28
Natural Language UnderstandingGLUE base (test dev)
CoLA MCC69.55
11
Subject-driven image generationDreamBooth
Fine-tuning Loss0.095
4
Showing 8 of 8 rows

Other info

Follow for update