Low-Shot Learning from Imaginary Data
About
Humans can quickly learn new visual concepts, perhaps because they can easily visualize or imagine what novel objects look like from different views. Incorporating this ability to hallucinate novel instances of new concepts might help machine vision systems perform better low-shot learning, i.e., learning concepts from few examples. We present a novel approach to low-shot learning that uses this idea. Our approach builds on recent progress in meta-learning ("learning to learn") by combining a meta-learner with a "hallucinator" that produces additional training examples, and optimizing both models jointly. Our hallucinator can be incorporated into a variety of meta-learners and provides significant gains: up to a 6 point boost in classification accuracy when only a single training example is available, yielding state-of-the-art performance on the challenging ImageNet low-shot classification benchmark.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | ImageNet 1K Challenge (novel classes) | Top-5 Acc77.4 | 110 | |
| Generalized Few-Shot Learning | ImageNet 2012 (Novel classes) | Top-5 Accuracy77.4 | 70 | |
| Few-shot Image Classification | ImageNet FS (novel) | Top-5 Acc0.814 | 59 | |
| Low-shot Image Classification | ImageNet 1k (novel classes) | Top-5 Acc81.4 | 57 | |
| Generalized Few-Shot Learning | ImageNet All classes 2012 | Top-5 Accuracy77.5 | 50 | |
| Few-shot Image Classification | ImageNet FS (all) | Top-5 Acc82.8 | 44 | |
| Image Classification | Imagenet FS (All classes) | Top-5 Acc78.7 | 30 | |
| Generalized Low-shot Classification | ImageNet Combined SEEN and UNSEEN | Top-5 Accuracy83.3 | 30 | |
| Image Classification | CIFAR-100 Long-Tailed (rho=50) (test) | -- | 18 | |
| Image Classification | CIFAR-10 Long-Tailed (rho=50) (test) | -- | 17 |