On First-Order Meta-Learning Algorithms
About
This paper considers meta-learning problems, where there is a distribution of tasks, and we would like to obtain an agent that performs well (i.e., learns quickly) when presented with a previously unseen task sampled from this distribution. We analyze a family of algorithms for learning a parameter initialization that can be fine-tuned quickly on a new task, using only first-order derivatives for the meta-learning updates. This family includes and generalizes first-order MAML, an approximation to MAML obtained by ignoring second-order derivatives. It also includes Reptile, a new algorithm that we introduce here, which works by repeatedly sampling a task, training on it, and moving the initialization towards the trained weights on that task. We expand on the results from Finn et al. showing that first-order meta-learning algorithms perform well on some well-established benchmarks for few-shot classification, and we provide theoretical analysis aimed at understanding why these algorithms work.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Few-shot classification | tieredImageNet (test) | Accuracy71.03 | 282 | |
| Few-shot Image Classification | Mini-Imagenet (test) | Accuracy65.99 | 235 | |
| 5-way Classification | miniImageNet (test) | -- | 231 | |
| Few-shot classification | Mini-ImageNet | 1-shot Acc49.97 | 175 | |
| Few-shot classification | miniImageNet standard (test) | 5-way 1-shot Acc49.97 | 138 | |
| 5-way Image Classification | tieredImageNet 5-way (test) | 1-shot Acc48.97 | 117 | |
| Few-shot classification | Omniglot (test) | Accuracy97.12 | 109 | |
| Few-shot classification | Mini-ImageNet 1-shot 5-way (test) | Accuracy49.97 | 82 | |
| Image Classification | Visual Decathlon Challenge 1.0 (test) | Mean Accuracy41.28 | 81 | |
| 5-way Image Classification | MiniImagenet | One-shot Accuracy49.97 | 67 |