Uncertainty in Model-Agnostic Meta-Learning using Variational Inference
About
We introduce a new, rigorously-formulated Bayesian meta-learning algorithm that learns a probability distribution of model parameter prior for few-shot learning. The proposed algorithm employs a gradient-based variational inference to infer the posterior of model parameters to a new task. Our algorithm can be applied to any model architecture and can be implemented in various machine learning paradigms, including regression and classification. We show that the models trained with our proposed meta-learning algorithm are well calibrated and accurate, with state-of-the-art calibration and classification results on two few-shot classification benchmarks (Omniglot and Mini-ImageNet), and competitive results in a multi-modal task-distribution regression.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Few-shot Image Classification | Mini-Imagenet (test) | Accuracy51.54 | 235 | |
| 5-way Image Classification | tieredImageNet 5-way (test) | 1-shot Acc69.87 | 117 | |
| 5-way Image Classification | MiniImagenet | One-shot Accuracy62.16 | 67 | |
| 5-way Image Classification | Mini-Imagenet (test) | -- | 46 | |
| 5-way Few-shot Classification | miniImageNet 5-way (test) | 1-shot Acc51.54 | 39 | |
| Classification | mini-ImageNet 5-shot (test) | Accuracy64.31 | 25 | |
| Few-shot classification | Omniglot standard (Random Split) | Accuracy (5-way 1-shot)98.43 | 9 | |
| Few-shot classification | Omniglot standard (Original Split) | Accuracy (5-way 1-shot)96.27 | 2 |