Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

LoRA-XS: Low-Rank Adaptation with Extremely Small Number of Parameters

About

The growth of large language models underscores the need for parameter-efficient fine-tuning. Despite its popularity, LoRA encounters storage and computational challenges when deploying multiple task- or user-specific modules. To address this, we introduce LoRA-XS, a novel fine-tuning method backed by a theoretical derivation. LoRA-XS drastically reduces trainable parameters by incorporating a small, trainable weight matrix between frozen low-rank matrices derived from the Singular Value Decomposition of pre-trained weights. This design enables LoRA-XS to reduce storage requirements by over 100x in 7B models compared to LoRA. Additionally, unlike other methods, LoRA-XS imposes no lower bound on trainable parameters - it can scale from a single parameter per module to arbitrarily large values, adapting to any storage or computational constraint. Evaluations on GLUE, GSM8K, MATH, and commonsense reasoning benchmarks across different model scales reveal that LoRA-XS consistently outperforms or matches LoRA and VeRA in accuracy, offering unmatched parameter efficiency. Our ablation studies highlight the significance of singular vectors in transformer weights, establishing LoRA-XS as a powerful, storage-efficient solution for scaling and personalizing large language models.

Klaudia Ba{\l}azy, Mohammadreza Banaei, Karl Aberer, Jacek Tabor• 2024

Related benchmarks

TaskDatasetResultRank
Mathematical ReasoningGSM8K (test)
Accuracy61.56
797
Mathematical ReasoningGSM8K (test)
Accuracy74.9
751
Question AnsweringARC Challenge
Accuracy84
749
Natural Language UnderstandingGLUE
SST-293.19
452
Mathematical ReasoningMATH (test)
Overall Accuracy31.28
433
Natural Language UnderstandingGLUE (test)
SST-2 Accuracy96.3
416
Question AnsweringARC Easy
Accuracy92.89
386
Reading ComprehensionBoolQ
Accuracy82.07
219
Commonsense ReasoningCommonsense Reasoning (BoolQ, PIQA, SIQA, HellaS., WinoG., ARC-e, ARC-c, OBQA) (test)
BoolQ Accuracy72.35
138
Image ClassificationVTAB 1k (test)--
121
Showing 10 of 18 rows

Other info

Follow for update