Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

A Simple Efficiency Incremental Learning Framework via Vision-Language Model with Nonlinear Multi-Adapters

About

Incremental Learning (IL) aims to learn new tasks while preserving previously acquired knowledge. Integrating the zero-shot learning capabilities of pre-trained vision-language models into IL methods has marked a significant advancement. However, these methods face three primary challenges: (1) the need for improved training efficiency; (2) reliance on a memory bank to store previous data; and (3) the necessity of a strong backbone to augment the model's capabilities. In this paper, we propose SimE, a Simple and Efficient framework that employs a vision-language model with adapters designed specifically for the IL task. We report a remarkable phenomenon: there is a nonlinear correlation between the number of adaptive adapter connections and the model's IL capabilities. While increasing adapter connections between transformer blocks improves model performance, adding more adaptive connections within transformer blocks during smaller incremental steps does not enhance, and may even degrade the model's IL ability. Extensive experimental results show that SimE surpasses traditional methods by 9.6% on TinyImageNet and outperforms other CLIP-based methods by 5.3% on CIFAR-100. Furthermore, we conduct a systematic study to enhance the utilization of the zero-shot capabilities of CLIP. We suggest replacing SimE's encoder with a CLIP model trained on larger datasets (e.g., LAION2B) and stronger architectures (e.g., ViT-L/14).

Haihua Luo, Xuming Ran, Jiangrong Shen, Timo H\"am\"al\"ainen, Zhonghua Chen, Qi Xu, Fengyu Cong• 2026

Related benchmarks

TaskDatasetResultRank
Class-incremental learningCIFAR-100
Averaged Incremental Accuracy85.94
248
Class-incremental learningImageNet-100
Avg Acc79.35
82
Class-incremental learningVTAB B0 Inc10
Last Accuracy78.04
38
Class-incremental learningImageNet-100 (10T)
Average Accuracy (A_T)89.77
35
Class-incremental learningCIFAR100 20 steps (test)
Last Accuracy86.64
33
Class-incremental learningCUB200 (100-20)
Avg Accuracy84.98
22
Class-incremental learningCIFAR100 50 steps (test)
Last Accuracy85.35
21
Class-incremental learningTinyImageNet 10 Steps (test)--
15
Class-incremental learningCIFAR100 10 Steps (test)
Average Accuracy91.66
12
Class-incremental learningTinyImageNet 5 Steps (test)
Average Accuracy86.47
12
Showing 10 of 16 rows

Other info

Follow for update