Meta-World: A Benchmark and Evaluation for Multi-Task and Meta Reinforcement Learning
About
Meta-reinforcement learning algorithms can enable robots to acquire new skills much more quickly, by leveraging prior experience to learn how to learn. However, much of the current research on meta-reinforcement learning focuses on task distributions that are very narrow. For example, a commonly used meta-reinforcement learning benchmark uses different running velocities for a simulated robot as different tasks. When policies are meta-trained on such narrow task distributions, they cannot possibly generalize to more quickly acquire entirely new tasks. Therefore, if the aim of these methods is to enable faster acquisition of entirely new behaviors, we must evaluate them on task distributions that are sufficiently broad to enable generalization to new behaviors. In this paper, we propose an open-source simulated benchmark for meta-reinforcement learning and multi-task learning consisting of 50 distinct robotic manipulation tasks. Our aim is to make it possible to develop algorithms that generalize to accelerate the acquisition of entirely new, held-out tasks. We evaluate 7 state-of-the-art meta-reinforcement learning and multi-task learning algorithms on these tasks. Surprisingly, while each task and its variations (e.g., with different object positions) can be learned with reasonable success, these algorithms struggle to learn with multiple tasks at the same time, even with as few as ten distinct training tasks. Our analysis and open-source environments pave the way for future research in multi-task learning and meta-learning that can enable meaningful generalization, thereby unlocking the full potential of these methods.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Multi-task reinforcement learning | Meta-World MT10 v1 (Fixed) | Success Rate88 | 12 | |
| Multi-task reinforcement learning | Meta-World MT10 v1 (train test) | Average Success61 | 9 | |
| Multi-task reinforcement learning | Meta-World MT50 v1 (Fixed) | Success Rate35.9 | 8 | |
| Task Diversity Assessment | RL Task Diversity Collections | Self-BLEU0.322 | 6 | |
| Multi-task reinforcement learning | Meta-World MT10 Conditioned v1 | Average Success Rate67.4 | 6 | |
| Multi-task reinforcement learning | Meta-World MT50 Conditioned v1 | Average Success Rate34.2 | 6 |