Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

On First-Order Meta-Learning Algorithms

About

This paper considers meta-learning problems, where there is a distribution of tasks, and we would like to obtain an agent that performs well (i.e., learns quickly) when presented with a previously unseen task sampled from this distribution. We analyze a family of algorithms for learning a parameter initialization that can be fine-tuned quickly on a new task, using only first-order derivatives for the meta-learning updates. This family includes and generalizes first-order MAML, an approximation to MAML obtained by ignoring second-order derivatives. It also includes Reptile, a new algorithm that we introduce here, which works by repeatedly sampling a task, training on it, and moving the initialization towards the trained weights on that task. We expand on the results from Finn et al. showing that first-order meta-learning algorithms perform well on some well-established benchmarks for few-shot classification, and we provide theoretical analysis aimed at understanding why these algorithms work.

Alex Nichol, Joshua Achiam, John Schulman• 2018

Related benchmarks

TaskDatasetResultRank
Few-shot classificationtieredImageNet (test)
Accuracy71.03
282
Few-shot Image ClassificationMini-Imagenet (test)
Accuracy65.99
235
5-way ClassificationminiImageNet (test)--
231
Few-shot classificationMini-ImageNet
1-shot Acc49.97
175
Few-shot classificationminiImageNet standard (test)
5-way 1-shot Acc49.97
138
5-way Image ClassificationtieredImageNet 5-way (test)
1-shot Acc48.97
117
Few-shot classificationOmniglot (test)
Accuracy97.12
109
Few-shot classificationMini-ImageNet 1-shot 5-way (test)
Accuracy49.97
82
Image ClassificationVisual Decathlon Challenge 1.0 (test)
Mean Accuracy41.28
81
5-way Image ClassificationMiniImagenet
One-shot Accuracy49.97
67
Showing 10 of 60 rows

Other info

Follow for update