Meta-Baseline: Exploring Simple Meta-Learning for Few-Shot Learning
About
Meta-learning has been the most common framework for few-shot learning in recent years. It learns the model from collections of few-shot classification tasks, which is believed to have a key advantage of making the training objective consistent with the testing objective. However, some recent works report that by training for whole-classification, i.e. classification on the whole label-set, it can get comparable or even better embedding than many meta-learning algorithms. The edge between these two lines of works has yet been underexplored, and the effectiveness of meta-learning in few-shot learning remains unclear. In this paper, we explore a simple process: meta-learning over a whole-classification pre-trained model on its evaluation metric. We observe this simple method achieves competitive performance to state-of-the-art methods on standard benchmarks. Our further analysis shed some light on understanding the trade-offs between the meta-learning objective and the whole-classification objective in few-shot learning.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Few-shot classification | tieredImageNet (test) | Accuracy83.29 | 282 | |
| Image Classification | MiniImagenet | Accuracy72.96 | 206 | |
| Few-shot classification | Mini-ImageNet | 1-shot Acc79.26 | 175 | |
| 5-way Few-shot Classification | MiniImagenet | Accuracy (5-shot)79.26 | 150 | |
| 5-way Few-shot Classification | Mini-Imagenet (test) | 1-shot Accuracy65.31 | 141 | |
| Few-shot classification | miniImageNet (test) | Accuracy79.26 | 120 | |
| Few-shot classification | Mini-Imagenet (test) | Accuracy79.26 | 113 | |
| Few-shot Image Classification | miniImageNet (test) | -- | 111 | |
| Few-shot classification | Mini-Imagenet 5-way 5-shot | Accuracy79.3 | 87 | |
| Few-shot Image Classification | tieredImageNet (test) | Accuracy83.74 | 86 |