Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Enhancing Online Continual Learning with Plug-and-Play State Space Model and Class-Conditional Mixture of Discretization

About

Online continual learning (OCL) seeks to learn new tasks from data streams that appear only once, while retaining knowledge of previously learned tasks. Most existing methods rely on replay, focusing on enhancing memory retention through regularization or distillation. However, they often overlook the adaptability of the model, limiting the ability to learn generalizable and discriminative features incrementally from online training data. To address this, we introduce a plug-and-play module, S6MOD, which can be integrated into most existing methods and directly improve adaptability. Specifically, S6MOD introduces an extra branch after the backbone, where a mixture of discretization selectively adjusts parameters in a selective state space model, enriching selective scan patterns such that the model can adaptively select the most sensitive discretization method for current dynamics. We further design a class-conditional routing algorithm for dynamic, uncertainty-based adjustment and implement a contrastive discretization loss to optimize it. Extensive experiments combining our module with various models demonstrate that S6MOD significantly enhances model adaptability, leading to substantial performance gains and achieving the state-of-the-art results.

Sihao Liu, Yibo Yang, Xiaojie Li, David A. Clifton, Bernard Ghanem• 2024

Related benchmarks

TaskDatasetResultRank
Online Continual LearningCIFAR-100 (test)
Accuracy56.32
42
Online Continual LearningTiny-ImageNet
Average Forgetting11.55
42
Online Continual LearningTiny ImageNet (test)
Avg Accuracy40.11
42
Online Continual LearningCIFAR10
Average Forgetting5.73
28
Online Continual LearningCIFAR10 (test)
Accuracy79.51
28
Showing 5 of 5 rows

Other info

Follow for update