Boosting Few-Shot Learning With Adaptive Margin Loss
About
Few-shot learning (FSL) has attracted increasing attention in recent years but remains challenging, due to the intrinsic difficulty in learning to generalize from a few examples. This paper proposes an adaptive margin principle to improve the generalization ability of metric-based meta-learning approaches for few-shot learning problems. Specifically, we first develop a class-relevant additive margin loss, where semantic similarity between each pair of classes is considered to separate samples in the feature embedding space from similar classes. Further, we incorporate the semantic context among all classes in a sampled training task and develop a task-relevant additive margin loss to better distinguish samples from different classes. Our adaptive margin method can be easily extended to a more realistic generalized FSL setting. Extensive experiments demonstrate that the proposed method can boost the performance of current metric-based meta-learning approaches, under both the standard FSL and generalized FSL settings.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| 5-way Few-shot Classification | Mini-Imagenet (test) | 1-shot Accuracy67.1 | 141 | |
| Few-shot classification | miniImageNet standard (test) | 5-way 1-shot Acc67.1 | 138 | |
| Few-shot classification | Mini-Imagenet (test) | Accuracy79.54 | 113 | |
| Few-shot Image Classification | miniImageNet (test) | -- | 111 | |
| Generalized Few-Shot Learning | ImageNet 2012 (Novel classes) | Top-5 Accuracy79.4 | 70 | |
| Generalized Few-Shot Learning | ImageNet All classes 2012 | Top-5 Accuracy80.2 | 50 | |
| 5-way 1-shot Classification | ImageNet mini | Top-1 Accuracy (ACC_1)60.31 | 31 | |
| Few-shot Image Classification | miniImageNet original (test) | 5-way 1-shot Acc67.1 | 30 | |
| 5-way 5-shot Classification | Mini-ImageNet | Mean Accuracy77.94 | 27 |