Causally Sufficient and Necessary Feature Expansion for Class-Incremental Learning
About
Current expansion-based methods for Class Incremental Learning (CIL) effectively mitigate catastrophic forgetting by freezing old features. However, such task-specific features learned from the new task may collide with the old features. From a causal perspective, spurious feature correlations are the main cause of this collision, manifesting in two scopes: (i) guided by empirical risk minimization (ERM), intra-task spurious correlations cause task-specific features to rely on shortcut features. These non-robust features are vulnerable to interference, inevitably drifting into the feature space of other tasks; (ii) inter-task spurious correlations induce semantic confusion between visually similar classes across tasks. To address this, we propose a Probability of Necessity and Sufficiency (PNS)-based regularization method to guide feature expansion in CIL. Specifically, we first extend the definition of PNS to expansion-based CIL, termed CPNS, which quantifies both the causal completeness of intra-task representations and the separability of inter-task representations. We then introduce a dual-scope counterfactual generator based on twin networks to ensure the measurement of CPNS, which simultaneously generates: (i) intra-task counterfactual features to minimize intra-task PNS risk and ensure causal completeness of task-specific features, and (ii) inter-task interfering features to minimize inter-task PNS risk, ensuring the separability of inter-task representations. Theoretical analyses confirm its reliability. The regularization is a plug-and-play method for expansion-based CIL to mitigate feature collision. Extensive experiments demonstrate the effectiveness of the proposed method.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Class-incremental learning | CIFAR-100 | Average Accuracy79.16 | 116 | |
| Class-incremental learning | ImageNet-100 (10T) | Average Accuracy (A_T)79.96 | 35 | |
| Class-incremental learning | CUB200 (100-20) | Avg Accuracy58.21 | 22 | |
| Class-incremental learning | CIFAR-100 50-10 scenario | Last Accuracy71.41 | 8 | |
| Class-incremental learning | ImageNet-100 50-10 scenario | Last Accuracy77.19 | 8 | |
| Class-incremental learning | ImageNet-1000 100-100 scenario | Last Accuracy62.44 | 8 |