SimSR: Simple Distance-based State Representation for Deep Reinforcement Learning
About
This work explores how to learn robust and generalizable state representation from image-based observations with deep reinforcement learning methods. Addressing the computational complexity, stringent assumptions and representation collapse challenges in existing work of bisimulation metric, we devise Simple State Representation (SimSR) operator. SimSR enables us to design a stochastic approximation method that can practically learn the mapping functions (encoders) from observations to latent representation space. In addition to the theoretical analysis and comparison with the existing work, we experimented and compared our work with recent state-of-the-art solutions in visual MuJoCo tasks. The results shows that our model generally achieves better performance and has better robustness and good generalization.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Continuous Control | DMControl 500k | Spin Score0.4 | 33 | |
| Visual Offline Reinforcement Learning | V-D4RL (various) | Cheetah-Run Medium391 | 8 | |
| C-SwingUp | DM_Control | Average Return854.8 | 6 | |
| W-Walk | DM_Control | Average Return929.9 | 6 | |
| F-Spin | DM_Control | Average Return962.4 | 6 | |
| BiC-Catch | DM_Control | Average Return938.8 | 6 | |
| C-SwingUpSparse | DM_Control | Average Return217.9 | 6 | |
| Continuous Control | DM_Control distraction setting (test) | BiC-Catch Score106.4 | 6 | |
| Ch-Run | DM Control | Average Return255.3 | 6 | |
| H-Stand | DM Control | Average Return6.2 | 6 |