Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Learning to Learn: Meta-Critic Networks for Sample Efficient Learning

About

We propose a novel and flexible approach to meta-learning for learning-to-learn from only a few examples. Our framework is motivated by actor-critic reinforcement learning, but can be applied to both reinforcement and supervised learning. The key idea is to learn a meta-critic: an action-value function neural network that learns to criticise any actor trying to solve any specified task. For supervised learning, this corresponds to the novel idea of a trainable task-parametrised loss generator. This meta-critic approach provides a route to knowledge transfer that can flexibly deal with few-shot and semi-supervised conditions for both reinforcement and supervised learning. Promising results are shown on both reinforcement and supervised learning problems.

Flood Sung, Li Zhang, Tao Xiang, Timothy Hospedales, Yongxin Yang• 2017

Related benchmarks

TaskDatasetResultRank
Mountain CarMountain Car held-out domains random mountain heights (test)
Avg. Failure Rate5
6
Reinforcement LearningCart-Pole OpenAI Gym (3 held-out domains (variable pole length and cart mass))
Return144.2
6
Reinforcement LearningCart-Pole Domain Generalization - Pole Length OpenAI Gym (3 held out domains)
Average Reward97.39
6
Showing 3 of 3 rows

Other info

Follow for update