Bayesian Meta-Learning for the Few-Shot Setting via Deep Kernels
About
Recently, different machine learning methods have been introduced to tackle the challenging few-shot learning scenario that is, learning from a small labeled dataset related to a specific task. Common approaches have taken the form of meta-learning: learning to learn on the new problem given the old. Following the recognition that meta-learning is implementing learning in a multi-level model, we present a Bayesian treatment for the meta-learning inner loop through the use of deep kernels. As a result we can learn a kernel that transfers to new tasks; we call this Deep Kernel Transfer (DKT). This approach has many advantages: is straightforward to implement as a single optimizer, provides uncertainty quantification, and does not require estimation of task-specific parameters. We empirically demonstrate that DKT outperforms several state-of-the-art algorithms in few-shot classification, and is the state of the art for cross-domain adaptation and regression. We conclude that complex meta-learning routines can be replaced by a simpler Bayesian model without loss of accuracy.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Few-shot Image Classification | Mini-Imagenet (test) | Accuracy49.73 | 235 | |
| 5-way Few-shot Classification | MiniImagenet | Accuracy (5-shot)62.85 | 150 | |
| Few-shot classification | CUB (test) | Accuracy63.46 | 145 | |
| 5-way Few-shot Classification | CUB | 5-shot Acc85.64 | 95 | |
| Classification | CUB (test) | Accuracy63.37 | 79 | |
| Few-shot classification | mini-ImageNet → CUB (test) | Accuracy (5-shot)57.23 | 75 | |
| Classification | Omniglot to EMNIST (test) | Accuracy90.3 | 51 | |
| 5-way Cross-domain Classification | mini-ImageNet -> CUB cross-domain (test) | Accuracy56.4 | 42 | |
| Few-shot classification | ImageNet mini (test) | ECE0.287 | 34 | |
| Classification | mini-ImageNet 5-shot (test) | Accuracy64 | 25 |