N$^2$M$^2$: Learning Navigation for Arbitrary Mobile Manipulation Motions in Unseen and Dynamic Environments
About
Despite its importance in both industrial and service robotics, mobile manipulation remains a significant challenge as it requires a seamless integration of end-effector trajectory generation with navigation skills as well as reasoning over long-horizons. Existing methods struggle to control the large configuration space, and to navigate dynamic and unknown environments. In previous work, we proposed to decompose mobile manipulation tasks into a simplified motion generator for the end-effector in task space and a trained reinforcement learning agent for the mobile base to account for kinematic feasibility of the motion. In this work, we introduce Neural Navigation for Mobile Manipulation (N$^2$M$^2$) which extends this decomposition to complex obstacle environments and enables it to tackle a broad range of tasks in real world settings. The resulting approach can perform unseen, long-horizon tasks in unexplored environments while instantly reacting to dynamic obstacles and environmental changes. At the same time, it provides a simple way to define new mobile manipulation tasks. We demonstrate the capabilities of our proposed approach in extensive simulation and real-world experiments on multiple kinematically diverse mobile manipulators. Code and videos are publicly available at http://mobile-rl.cs.uni-freiburg.de.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Mobile Manipulation | cross-room experimental configuration 1.0 (test) | TCR29 | 7 | |
| Mobile Manipulation | Home-scene Cross-room transfer v1 | AIKF15.1 | 5 | |
| Mobile Manipulation | Warehouse Cross-room transfer v1 | AIKF14.3 | 5 | |
| Mobile Manipulation | Dynamic-scene Cross-room transfer v1 | AIKF18.7 | 5 | |
| Mobile Manipulation | Home-scene | AIKF11 | 4 | |
| Mobile Manipulation | Warehouse | AIKF Score7.1 | 4 | |
| Mobile Manipulation | Dynamic Scene | AIKF16.5 | 4 |