Beyond OOD State Actions: Supported Cross-Domain Offline Reinforcement Learning
About
Offline reinforcement learning (RL) aims to learn a policy using only pre-collected and fixed data. Although avoiding the time-consuming online interactions in RL, it poses challenges for out-of-distribution (OOD) state actions and often suffers from data inefficiency for training. Despite many efforts being devoted to addressing OOD state actions, the latter (data inefficiency) receives little attention in offline RL. To address this, this paper proposes the cross-domain offline RL, which assumes offline data incorporate additional source-domain data from varying transition dynamics (environments), and expects it to contribute to the offline data efficiency. To do so, we identify a new challenge of OOD transition dynamics, beyond the common OOD state actions issue, when utilizing cross-domain offline data. Then, we propose our method BOSA, which employs two support-constrained objectives to address the above OOD issues. Through extensive experiments in the cross-domain offline RL setting, we demonstrate BOSA can greatly improve offline data efficiency: using only 10\% of the target data, BOSA could achieve {74.4\%} of the SOTA offline RL performance that uses 100\% of the target data. Additionally, we also show BOSA can be effortlessly plugged into model-based offline RL and noising data augmentation techniques (used for generating source-domain data), which naturally avoids the potential dynamics mismatch between target-domain data and newly generated source-domain data.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Offline Reinforcement Learning | hopper medium | Normalized Score20.6 | 52 | |
| Offline Reinforcement Learning | walker2d medium | Normalized Score10.6 | 51 | |
| Offline Reinforcement Learning | walker2d medium-replay | Normalized Score0.00e+0 | 50 | |
| Offline Reinforcement Learning | hopper medium-replay | Normalized Score3.7 | 44 | |
| Offline Reinforcement Learning | halfcheetah medium | Normalized Score38.9 | 43 | |
| Offline Reinforcement Learning | halfcheetah medium-replay | Normalized Score20 | 43 | |
| Offline Reinforcement Learning | Antmaze Medium play offline (target domain) | Target Domain Score (Normalized)158.4 | 42 | |
| Locomotion | D4RL HalfCheetah medium-offline | Normalized Score27.36 | 36 | |
| Locomotion | D4RL Hopper medium-offline | Score13.97 | 36 | |
| Locomotion | D4RL Ant medium-offline | Normalized Score55.51 | 36 |