Prototype Rectification for Few-Shot Learning
About
Few-shot learning requires to recognize novel classes with scarce labeled data. Prototypical network is useful in existing researches, however, training on narrow-size distribution of scarce data usually tends to get biased prototypes. In this paper, we figure out two key influencing factors of the process: the intra-class bias and the cross-class bias. We then propose a simple yet effective approach for prototype rectification in transductive setting. The approach utilizes label propagation to diminish the intra-class bias and feature shifting to diminish the cross-class bias. We also conduct theoretical analysis to derive its rationality as well as the lower bound of the performance. Effectiveness is shown on three few-shot benchmarks. Notably, our approach achieves state-of-the-art performance on both miniImageNet (70.31% on 1-shot and 81.89% on 5-shot) and tieredImageNet (78.74% on 1-shot and 86.92% on 5-shot).
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Few-shot classification | tieredImageNet (test) | Accuracy86.92 | 282 | |
| Few-shot Image Classification | Mini-Imagenet (test) | Accuracy81.89 | 235 | |
| Image Classification | ImageNet | Accuracy61.7 | 184 | |
| Few-shot classification | Mini-ImageNet | 1-shot Acc70.3 | 175 | |
| 5-way Few-shot Classification | MiniImagenet | Accuracy (5-shot)84.14 | 150 | |
| Few-shot classification | miniImageNet standard (test) | -- | 138 | |
| 5-way Few-shot Classification | CUB | 5-shot Acc88.7 | 95 | |
| Few-shot Image Classification | tieredImageNet | -- | 90 | |
| Image Classification | Mini-Imagenet (test) | Acc (5-shot)51.1 | 75 | |
| Few-shot Image Classification | mini-ImageNet K=20 (test) | Accuracy58.4 | 56 |