Proto Successor Measure: Representing the Behavior Space of an RL Agent
About
Having explored an environment, intelligent agents should be able to transfer their knowledge to most downstream tasks within that environment without additional interactions. Referred to as "zero-shot learning", this ability remains elusive for general-purpose reinforcement learning algorithms. While recent works have attempted to produce zero-shot RL agents, they make assumptions about the nature of the tasks or the structure of the MDP. We present Proto Successor Measure: the basis set for all possible behaviors of a Reinforcement Learning Agent in a dynamical system. We prove that any possible behavior (represented using visitation distributions) can be represented using an affine combination of these policy-independent basis functions. Given a reward function at test time, we simply need to find the right set of linear weights to combine these bases corresponding to the optimal policy. We derive a practical algorithm to learn these basis functions using reward-free interaction data from the environment and show that our approach can produce the optimal policy at test time for any given reward function without additional environmental interactions. Project page: https://agarwalsiddhant10.github.io/projects/psm.html.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Offline Reinforcement Learning | halfcheetah medium v2 | Average Score42.64 | 27 | |
| Offline Reinforcement Learning | halfcheetah medium-expert v2 | Normalized Score49.92 | 18 | |
| Offline Reinforcement Learning | walker2d medium v2 | Normalized Score55.7 | 18 | |
| Offline Reinforcement Learning | hopper medium v2 | -- | 14 | |
| Offline Reinforcement Learning | Walker2d Medium-Expert v2 | Average Score79.32 | 7 | |
| Offline Reinforcement Learning | Hopper Medium-Expert v2 | Average Score14.59 | 7 | |
| Offline Reinforcement Learning | DeepMind Control Suite Cheetah (test) | Run Score244.4 | 5 | |
| Offline Reinforcement Learning | DeepMind Control Suite Quadruped (test) | Stand Score842.9 | 5 | |
| Offline Reinforcement Learning | DeepMind Control Suite Walker (test) | Stand Score872.6 | 5 | |
| Offline Reinforcement Learning | Pointmass DeepMind Control Suite (test) | Performance (Top Left)831.4 | 5 |