Dream to Control: Learning Behaviors by Latent Imagination
About
Learned world models summarize an agent's experience to facilitate learning complex behaviors. While learning world models from high-dimensional sensory inputs is becoming feasible through deep learning, there are many potential ways for deriving behaviors from them. We present Dreamer, a reinforcement learning agent that solves long-horizon tasks from images purely by latent imagination. We efficiently learn behaviors by propagating analytic gradients of learned state values back through trajectories imagined in the compact state space of a learned world model. On 20 challenging visual control tasks, Dreamer exceeds existing approaches in data-efficiency, computation time, and final performance.
Danijar Hafner, Timothy Lillicrap, Jimmy Ba, Mohammad Norouzi• 2019
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Continuous Control | DMControl 500k | Spin Score796 | 33 | |
| Continuous Control | DMControl 100k | DMControl: Finger Spin Score498.9 | 29 | |
| Visual Reinforcement Learning | DMControl Reacher Easy | Episode Return314 | 16 | |
| Visual Reinforcement Learning | DMControl Walker Walk | Episode Return277 | 16 | |
| Visual Reinforcement Learning | DMControl Cheetah Run | Episode Return235 | 16 | |
| Visual Reinforcement Learning | DMControl Ball in cup, Catch | Episode Return246 | 16 | |
| Visual Reinforcement Learning | DMControl Finger, Spin | Episode Return341 | 16 | |
| Visual Reinforcement Learning | DMControl Cartpole, Swingup | Episode Return326 | 16 | |
| Pixel-based Control | DeepMind Control Suite 500k environment steps | Cheetah Run Score570 | 9 | |
| Pixel-based Control | DeepMind Control Suite 100k steps | Cheetah/Run Score235 | 9 |
Showing 10 of 39 rows