Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

InfLoRA: Interference-Free Low-Rank Adaptation for Continual Learning

About

Continual learning requires the model to learn multiple tasks sequentially. In continual learning, the model should possess the ability to maintain its performance on old tasks (stability) and the ability to adapt to new tasks continuously (plasticity). Recently, parameter-efficient fine-tuning (PEFT), which involves freezing a pre-trained model and injecting a small number of learnable parameters to adapt to downstream tasks, has gained increasing popularity in continual learning. Although existing continual learning methods based on PEFT have demonstrated superior performance compared to those not based on PEFT, most of them do not consider how to eliminate the interference of the new task on the old tasks, which inhibits the model from making a good trade-off between stability and plasticity. In this work, we propose a new PEFT method, called interference-free low-rank adaptation (InfLoRA), for continual learning. InfLoRA injects a small number of parameters to reparameterize the pre-trained weights and shows that fine-tuning these injected parameters is equivalent to fine-tuning the pre-trained weights within a subspace. Furthermore, InfLoRA designs this subspace to eliminate the interference of the new task on the old tasks, making a good trade-off between stability and plasticity. Experimental results show that InfLoRA outperforms existing state-of-the-art continual learning methods on multiple datasets.

Yan-Shuo Liang, Wu-Jun Li• 2024

Related benchmarks

TaskDatasetResultRank
Class-incremental learningCIFAR-100
Averaged Incremental Accuracy91.71
234
Class-incremental learningImageNet A
Average Accuracy56.84
86
Class-incremental learningCIFAR-100 10 (test)
Average Top-1 Accuracy91.7
75
Class-incremental learningSplit ImageNet-R
Average Forgetting Measure5.73
57
Continual LearningStandard CL Benchmark
Avg Final Acc0.796
50
Continual LearningLarge Number of Tasks
Average Performance69.8
50
Class-incremental learningImageNet-R 10-task
FAA81.38
44
Class-incremental learningImageNet-R 20-task
Average Accuracy78.87
33
Continual LearningCIFAR100 (test)
Mean Accuracy91.33
31
Class-incremental learningVTAB
Avg Accuracy89.61
31
Showing 10 of 46 rows

Other info

Code

Follow for update