Probabilistic Model-Agnostic Meta-Learning
About
Meta-learning for few-shot learning entails acquiring a prior over previous tasks and experiences, such that new tasks be learned from small amounts of data. However, a critical challenge in few-shot learning is task ambiguity: even when a powerful prior can be meta-learned from a large number of prior tasks, a small dataset for a new task can simply be too ambiguous to acquire a single model (e.g., a classifier) for that task that is accurate. In this paper, we propose a probabilistic meta-learning algorithm that can sample models for a new task from a model distribution. Our approach extends model-agnostic meta-learning, which adapts to new tasks via gradient descent, to incorporate a parameter distribution that is trained via a variational lower bound. At meta-test time, our algorithm adapts via a simple procedure that injects noise into gradient descent, and at meta-training time, the model is trained such that this stochastic adaptation procedure produces samples from the approximate model posterior. Our experimental results show that our method can sample plausible classifiers and regressors in ambiguous few-shot learning problems. We also show how reasoning about ambiguity can also be used for downstream active learning problems.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Few-shot Image Classification | Mini-Imagenet (test) | Accuracy50.13 | 235 | |
| 5-way Classification | miniImageNet (test) | Accuracy50.13 | 231 | |
| Few-shot classification | miniImageNet standard (test) | 5-way 1-shot Acc50.13 | 138 | |
| 5-way Few-shot Classification | miniImageNet standard (test) | Accuracy50.13 | 91 | |
| 5-way Image Classification | MiniImagenet | One-shot Accuracy50.13 | 67 | |
| Image Classification | miniImageNet standard (test) | Accuracy50.13 | 61 | |
| 5-way Image Classification | Mini-Imagenet (test) | -- | 46 | |
| 5-way Few-shot Classification | miniImageNet 5-way (test) | 1-shot Acc50.13 | 39 |