Self-Sustaining Representation Expansion for Non-Exemplar Class-Incremental Learning
About
Non-exemplar class-incremental learning is to recognize both the old and new classes when old class samples cannot be saved. It is a challenging task since representation optimization and feature retention can only be achieved under supervision from new classes. To address this problem, we propose a novel self-sustaining representation expansion scheme. Our scheme consists of a structure reorganization strategy that fuses main-branch expansion and side-branch updating to maintain the old features, and a main-branch distillation scheme to transfer the invariant knowledge. Furthermore, a prototype selection mechanism is proposed to enhance the discrimination between the old and new classes by selectively incorporating new samples into the distillation process. Extensive experiments on three benchmarks demonstrate significant incremental performance, outperforming the state-of-the-art methods by a margin of 3%, 3% and 6%, respectively.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Incremental Learning | TinyImageNet | Avg Incremental Accuracy50.39 | 83 | |
| Class-incremental learning | CIFAR100 (test) | -- | 76 | |
| Incremental Learning | ImageNet subset | Average Accuracy67.69 | 58 | |
| Incremental Learning | CIFAR-100 | Average Accuracy65.88 | 51 | |
| Exemplar-Free Class-Incremental Learning | CIFAR-100 | Avg Top-1 Inc Acc65.9 | 38 | |
| Exemplar-Free Class-Incremental Learning | TinyImageNet | Top-1 Acc (Inc)50.4 | 32 | |
| Exemplar-Free Class-Incremental Learning | CIFAR-100 (test) | Accuracy Last (Alast)47.3 | 30 | |
| Exemplar-Free Class-Incremental Learning | ImageNet subset (test) | A_last43.8 | 30 | |
| Class-incremental learning | TinyImageNet (test) | Accuracy50.39 | 29 | |
| Incremental Learning | CIFAR-100 (test) | Accuracy (S9)55.7 | 26 |