Consistency Models as a Rich and Efficient Policy Class for Reinforcement Learning
About
Score-based generative models like the diffusion model have been testified to be effective in modeling multi-modal data from image generation to reinforcement learning (RL). However, the inference process of diffusion model can be slow, which hinders its usage in RL with iterative sampling. We propose to apply the consistency model as an efficient yet expressive policy representation, namely consistency policy, with an actor-critic style algorithm for three typical RL settings: offline, offline-to-online and online. For offline RL, we demonstrate the expressiveness of generative models as policies from multi-modal data. For offline-to-online RL, the consistency policy is shown to be more computational efficient than diffusion policy, with a comparable performance. For online RL, the consistency policy demonstrates significant speedup and even higher average performances than the diffusion policy.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Offline Reinforcement Learning | D4RL antmaze-umaze (diverse) | Normalized Score77.6 | 40 | |
| Offline Reinforcement Learning | D4RL MuJoCo Hopper medium standard | Normalized Score80.7 | 36 | |
| Offline Reinforcement Learning | D4RL Adroit pen (human) | Normalized Return64 | 32 | |
| Offline Reinforcement Learning | D4RL Adroit pen (cloned) | Normalized Return56 | 32 | |
| Offline Reinforcement Learning | MuJoCo hopper D4RL (medium-replay) | Normalized Return99.7 | 26 | |
| Offline Reinforcement Learning | D4RL antmaze-large (play) | Normalized Score0.00e+0 | 26 | |
| Offline Reinforcement Learning | D4RL antmaze-med (diverse) | Normalized Score0.00e+0 | 26 | |
| Offline Reinforcement Learning | D4RL antmaze-large (diverse) | Normalized Score0.00e+0 | 26 | |
| Offline Reinforcement Learning | D4RL Adroit hammer-human | Normalized Score200 | 22 | |
| Offline Reinforcement Learning | D4RL Adroit door-human | Mean Normalized Score5 | 22 |