SAPG: Split and Aggregate Policy Gradients
About
Despite extreme sample inefficiency, on-policy reinforcement learning, aka policy gradients, has become a fundamental tool in decision-making problems. With the recent advances in GPU-driven simulation, the ability to collect large amounts of data for RL training has scaled exponentially. However, we show that current RL methods, e.g. PPO, fail to ingest the benefit of parallelized environments beyond a certain point and their performance saturates. To address this, we propose a new on-policy RL algorithm that can effectively leverage large-scale environments by splitting them into chunks and fusing them back together via importance sampling. Our algorithm, termed SAPG, shows significantly higher performance across a variety of challenging environments where vanilla PPO and other strong baselines fail to achieve high performance. Website at https://sapg-rl.github.io/
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Dexterous Manipulation | Throw | Average Performance22.51 | 4 | |
| Dexterous Manipulation | ShadowHand | Average Performance1.29e+4 | 4 | |
| Dexterous Manipulation | Regrasping | Average Performance37.2 | 4 | |
| Dexterous Manipulation | Reorientation | Average Performance38.79 | 4 | |
| Dexterous Manipulation | AllegroHand | Average Performance1.20e+4 | 4 | |
| Dexterous Manipulation | Two-Arms Reorientation | Average Performance5.11 | 4 |