Causal Navigation by Continuous-time Neural Networks
About
Imitation learning enables high-fidelity, vision-based learning of policies within rich, photorealistic environments. However, such techniques often rely on traditional discrete-time neural models and face difficulties in generalizing to domain shifts by failing to account for the causal relationships between the agent and the environment. In this paper, we propose a theoretical and experimental framework for learning causal representations using continuous-time neural networks, specifically over their discrete-time counterparts. We evaluate our method in the context of visual-control learning of drones over a series of complex tasks, ranging from short- and long-term navigation, to chasing static and dynamic objects through photorealistic environments. Our results demonstrate that causal continuous-time deep models can perform robust navigation tasks, where advanced recurrent models fail. These models learn complex causal control representations directly from raw visual inputs and scale to solve a variety of tasks using imitation learning.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Hiking | RedWood Forest (val) | Cosine Similarity Loss-0.859 | 8 | |
| Visual Drone Navigation (Chasing) | AirSim Chasing (test) | Success Rate (Clear)78 | 5 | |
| Visual Drone Navigation (Hiking) | AirSim Hiking (test) | Success Rate (Clear)3.00e+3 | 5 | |
| Visual Drone Navigation (Static Target) | AirSim Static Target (test) | Success Rate (Clear)0.48 | 5 | |
| Chasing objects | Neighborhood (val) | Cosine Similarity Loss-0.975 | 4 | |
| Hiking | Neighborhood (val) | Cosine Similarity Loss-0.711 | 4 | |
| Short-term navigation | Neighborhood (val) | Cosine Similarity Loss-0.855 | 4 | |
| Chasing objects | RedWood Forest (val) | Cosine Similarity Loss-0.936 | 4 |