Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

VeRA: Vector-based Random Matrix Adaptation

About

Low-rank adapation (LoRA) is a popular method that reduces the number of trainable parameters when finetuning large language models, but still faces acute storage challenges when scaling to even larger models or deploying numerous per-user or per-task adapted models. In this work, we present Vector-based Random Matrix Adaptation (VeRA), which significantly reduces the number of trainable parameters compared to LoRA, yet maintains the same performance. It achieves this by using a single pair of low-rank matrices shared across all layers and learning small scaling vectors instead. We demonstrate its effectiveness on the GLUE and E2E benchmarks, image classification tasks, and show its application in instruction-tuning of 7B and 13B language models.

Dawid J. Kopiczko, Tijmen Blankevoort, Yuki M. Asano• 2023

Related benchmarks

TaskDatasetResultRank
Question AnsweringARC Challenge
Accuracy81.02
906
Commonsense ReasoningPIQA
Accuracy78.63
751
Question AnsweringARC Easy
Accuracy94.05
597
Natural Language UnderstandingGLUE
SST-293.89
531
Natural Language UnderstandingGLUE (test)
SST-2 Accuracy96.1
416
Commonsense ReasoningCommon Sense Reasoning Tasks
Avg Score67.7
316
Reading ComprehensionRACE high
Accuracy78.84
295
Image ClassificationGTSRB
Accuracy91.23
291
Reading ComprehensionBoolQ
Accuracy85.85
279
Image ClassificationVTAB 1K
Overall Mean Accuracy69.9
258
Showing 10 of 40 rows

Other info

Follow for update