CALF: Communication-Aware Learning Framework for Distributed Reinforcement Learning
About
Distributed reinforcement learning policies face network delays, jitter, and packet loss when deployed across edge devices and cloud servers. Standard RL training assumes zero-latency interaction, causing severe performance degradation under realistic network conditions. We introduce CALF (Communication-Aware Learning Framework), which trains policies under realistic network models during simulation. Systematic experiments demonstrate that network-aware training substantially reduces deployment performance gaps compared to network-agnostic baselines. Distributed policy deployments across heterogeneous hardware validate that explicitly modelling communication constraints during training enables robust real-world execution. These findings establish network conditions as a major axis of sim-to-real transfer for Wi-Fi-like distributed deployments, complementing physics and visual domain randomisation.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Goal Reaching | MiniGrid Real-Eth | Success Rate85 | 3 | |
| Goal Reaching | MiniGrid Wi-Fi-N | Success Rate81 | 3 | |
| Goal Reaching | MiniGrid Wi-Fi-D | Success Rate74 | 3 | |
| Pole Balancing | CartPole Sim+Net | Mean Episode Return472 | 3 | |
| Pole Balancing | CartPole Real-Eth | Mean Episode Return458 | 3 | |
| Pole Balancing | CartPole Wi-Fi-N | Mean Episode Return442 | 3 | |
| Pole Balancing | CartPole Wi-Fi-D | Mean Episode Return378 | 3 | |
| Goal Reaching | MiniGrid Sim+Net | Success Rate87 | 3 | |
| Goal Reaching | MiniGrid Sim-Clean | Success Rate89 | 3 | |
| Pole Balancing | CartPole Sim-Clean | Mean Episode Return476 | 3 |