Power Normalizing Second-order Similarity Network for Few-shot Learning
About
Second- and higher-order statistics of data points have played an important role in advancing the state of the art on several computer vision problems such as the fine-grained image and scene recognition. However, these statistics need to be passed via an appropriate pooling scheme to obtain the best performance. Power Normalizations are non-linear activation units which enjoy probability-inspired derivations and can be applied in CNNs. In this paper, we propose a similarity learning network leveraging second-order information and Power Normalizations. To this end, we propose several formulations capturing second-order statistics and derive a sigmoid-like Power Normalizing function to demonstrate its interpretability. Our model is trained end-to-end to learn the similarity between the support set and query images for the problem of one- and few-shot learning. The evaluations on Omniglot, miniImagenet and Open MIC datasets demonstrate that this network obtains state-of-the-art results on several few-shot learning protocols.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| 5-way Few-shot Classification | MiniImagenet | Accuracy (5-shot)50.95 | 150 | |
| 5-way Few-shot Image Classification | CUB-200 2011 (meta-test) | 1-shot Acc46.72 | 16 | |
| Few-shot classification | VGG Flower Meta-Dataset 1.0 (test) | Accuracy (1-shot)71.9 | 16 | |
| Image Classification | tiered-Imagenet novel (test) | Top-1 Acc (1-shot)58.62 | 10 | |
| 5-way 1-shot learning | Open MIC Protocol I | Pairwise Accuracy (p1 to p2)78.6 | 8 |