Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Power Normalizing Second-order Similarity Network for Few-shot Learning

About

Second- and higher-order statistics of data points have played an important role in advancing the state of the art on several computer vision problems such as the fine-grained image and scene recognition. However, these statistics need to be passed via an appropriate pooling scheme to obtain the best performance. Power Normalizations are non-linear activation units which enjoy probability-inspired derivations and can be applied in CNNs. In this paper, we propose a similarity learning network leveraging second-order information and Power Normalizations. To this end, we propose several formulations capturing second-order statistics and derive a sigmoid-like Power Normalizing function to demonstrate its interpretability. Our model is trained end-to-end to learn the similarity between the support set and query images for the problem of one- and few-shot learning. The evaluations on Omniglot, miniImagenet and Open MIC datasets demonstrate that this network obtains state-of-the-art results on several few-shot learning protocols.

Hongguang Zhang, Piotr Koniusz• 2018

Related benchmarks

TaskDatasetResultRank
5-way Few-shot ClassificationMiniImagenet
Accuracy (5-shot)50.95
150
5-way Few-shot Image ClassificationCUB-200 2011 (meta-test)
1-shot Acc46.72
16
Few-shot classificationVGG Flower Meta-Dataset 1.0 (test)
Accuracy (1-shot)71.9
16
Image Classificationtiered-Imagenet novel (test)
Top-1 Acc (1-shot)58.62
10
5-way 1-shot learningOpen MIC Protocol I
Pairwise Accuracy (p1 to p2)78.6
8
Showing 5 of 5 rows

Other info

Follow for update