Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Parameter-Efficient Fine-Tuning via Circular Convolution

About

Low-Rank Adaptation (LoRA) has gained popularity for fine-tuning large foundation models, leveraging low-rank matrices $\mathbf{A}$ and $\mathbf{B}$ to represent weight changes (i.e., $\Delta \mathbf{W} = \mathbf{B} \mathbf{A}$). This method reduces trainable parameters and mitigates heavy memory consumption associated with full delta matrices by sequentially multiplying $\mathbf{A}$ and $\mathbf{B}$ with the activation. Despite its success, the intrinsic low-rank characteristic may limit its performance. Although several variants have been proposed to address this issue, they often overlook the crucial computational and memory efficiency brought by LoRA. In this paper, we propose Circular Convolution Adaptation (C$^3$A), which not only achieves high-rank adaptation with enhanced performance but also excels in both computational power and memory utilization. Extensive experiments demonstrate that C$^3$A consistently outperforms LoRA and its variants across various fine-tuning tasks.

Aochuan Chen, Jiashun Cheng, Zijing Liu, Ziqi Gao, Fugee Tsung, Yu Li, Jia Li• 2024

Related benchmarks

TaskDatasetResultRank
Natural Language UnderstandingGLUE (test)
SST-2 Accuracy95.78
416
Image ClassificationDTD (test)
Accuracy82.62
181
Commonsense ReasoningCommonsense Reasoning (BoolQ, PIQA, SIQA, HellaS., WinoG., ARC-e, ARC-c, OBQA) (test)
BoolQ Accuracy76.9
138
Image ClassificationEuroSAT (test)
Accuracy98.88
59
Image ClassificationCars (test)
Accuracy84.94
57
Image ClassificationPets (test)
Accuracy94.48
36
Code GenerationHumanEval and MBPP
Overall Average Score58.7
30
Mathematical ReasoningGSM8K and MATH
GSM8K Score78.4
27
Image ClassificationFGVC (test)
Accuracy63.8
25
Image ClassificationRESISC (test)
Accuracy95.94
17
Showing 10 of 10 rows

Other info

Follow for update