Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Vector Quantization Prompting for Continual Learning

About

Continual learning requires to overcome catastrophic forgetting when training a single model on a sequence of tasks. Recent top-performing approaches are prompt-based methods that utilize a set of learnable parameters (i.e., prompts) to encode task knowledge, from which appropriate ones are selected to guide the fixed pre-trained model in generating features tailored to a certain task. However, existing methods rely on predicting prompt identities for prompt selection, where the identity prediction process cannot be optimized with task loss. This limitation leads to sub-optimal prompt selection and inadequate adaptation of pre-trained features for a specific task. Previous efforts have tried to address this by directly generating prompts from input queries instead of selecting from a set of candidates. However, these prompts are continuous, which lack sufficient abstraction for task knowledge representation, making them less effective for continual learning. To address these challenges, we propose VQ-Prompt, a prompt-based continual learning method that incorporates Vector Quantization (VQ) into end-to-end training of a set of discrete prompts. In this way, VQ-Prompt can optimize the prompt selection process with task loss and meanwhile achieve effective abstraction of task knowledge for continual learning. Extensive experiments show that VQ-Prompt outperforms state-of-the-art continual learning methods across a variety of benchmarks under the challenging class-incremental setting. The code is available at \href{https://github.com/jiaolifengmi/VQ-Prompt}{this https URL}.

Li Jiao, Qiuxia Lai, Yu Li, Qiang Xu• 2024

Related benchmarks

TaskDatasetResultRank
Class-incremental learningImageNet-R 10-task
FAA78.71
44
Class-incremental learningSplit CIFAR-100 (10-task)
CAA92.84
41
Class-incremental learningImageNet-R 20-task--
33
Class-incremental learningImageNet-R 5-task--
27
Class-incremental learningCUB-200 Split (10-task)
FAA86.72
10
Continual LearningImageNet-A (20 tasks)
FAA52.96
2
Continual LearningVTAB five 10-class datasets
FAA90.46
2
Showing 7 of 7 rows

Other info

Code

Follow for update