Self-Promoted Prototype Refinement for Few-Shot Class-Incremental Learning
About
Few-shot class-incremental learning is to recognize the new classes given few samples and not forget the old classes. It is a challenging task since representation optimization and prototype reorganization can only be achieved under little supervision. To address this problem, we propose a novel incremental prototype learning scheme. Our scheme consists of a random episode selection strategy that adapts the feature representation to various generated incremental episodes to enhance the corresponding extensibility, and a self-promoted prototype refinement mechanism which strengthens the expression ability of the new classes by explicitly considering the dependencies among different classes. Particularly, a dynamic relation projection module is proposed to calculate the relation matrix in a shared embedding space and leverage it as the factor for bootstrapping the update of prototypes. Extensive experiments on three benchmark datasets demonstrate the above-par incremental performance, outperforming state-of-the-art methods by a margin of 13%, 17% and 11%, respectively.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Few-Shot Class-Incremental Learning | miniImageNet (test) | Accuracy (Session 1)74 | 173 | |
| Few-Shot Class-Incremental Learning | CIFAR100 (test) | Session 4 Top-1 Acc53.69 | 122 | |
| Few-Shot Class-Incremental Learning | CUB200 (test) | Accuracy (Session 1)68.68 | 92 | |
| Few-Shot Class-Incremental Learning | CUB-200 | Session 1 Accuracy61.85 | 75 | |
| Few-Shot Class-Incremental Learning | CUB200 (incremental sessions) | Session 0 Accuracy68.68 | 37 | |
| Few-Shot Class-Incremental Learning | miniImageNet 5-way 5-shot (incremental) | Accuracy S061.45 | 21 | |
| Few-Shot Class-Incremental Learning | CUB-200 | Session 0 Accuracy68.68 | 21 | |
| Few-Shot Class-Incremental Learning | miniImageNet, CUB-200, and CIFAR-100 Average (test) | Base Accuracy64.74 | 18 |