Leveraging the Feature Distribution in Transfer-based Few-Shot Learning
About
Few-shot classification is a challenging problem due to the uncertainty caused by using few labelled samples. In the past few years, many methods have been proposed to solve few-shot classification, among which transfer-based methods have proved to achieve the best performance. Following this vein, in this paper we propose a novel transfer-based method that builds on two successive steps: 1) preprocessing the feature vectors so that they become closer to Gaussian-like distributions, and 2) leveraging this preprocessing using an optimal-transport inspired algorithm (in the case of transductive settings). Using standardized vision benchmarks, we prove the ability of the proposed methodology to achieve state-of-the-art accuracy with various datasets, backbone architectures and few-shot settings.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Few-shot classification | tieredImageNet (test) | -- | 282 | |
| Image Classification | CUB | Accuracy91.55 | 249 | |
| Few-shot Image Classification | Mini-Imagenet (test) | -- | 235 | |
| Image Classification | ImageNet | Accuracy64 | 184 | |
| Few-shot classification | Mini-ImageNet | 1-shot Acc82.9 | 175 | |
| Few-shot Image Classification | miniImageNet (test) | -- | 111 | |
| Few-shot classification | Mini-Imagenet 5-way 5-shot | Accuracy88.8 | 87 | |
| Few-shot Image Classification | tieredImageNet (test) | -- | 86 | |
| Few-shot classification | Mini-ImageNet 1-shot 5-way (test) | Accuracy82.92 | 82 | |
| Few-shot classification | mini-ImageNet → CUB (test) | Accuracy (5-shot)76.51 | 75 |