SplitNet: Sim2Sim and Task2Task Transfer for Embodied Visual Navigation
About
We propose SplitNet, a method for decoupling visual perception and policy learning. By incorporating auxiliary tasks and selective learning of portions of the model, we explicitly decompose the learning objectives for visual navigation into perceiving the world and acting on that perception. We show dramatic improvements over baseline models on transferring between simulators, an encouraging step towards Sim2Real. Additionally, SplitNet generalizes better to unseen environments from the same simulator and transfers faster and more effectively to novel embodied navigation tasks. Further, given only a small sample from a target domain, SplitNet can match the performance of traditional end-to-end pipelines which receive the entire dataset. Code is available https://github.com/facebookresearch/splitnet
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| ObjectNav (Label goal) | Gibson tiny (test) | Success Rate9 | 20 | |
| ObjectNav (Audio goal) | Gibson tiny (test) | Success Rate8.8 | 10 | |
| ObjectNav (Audio) | Gibson (test) | Success Rate8.8 | 10 | |
| ObjectNav (Label) | Gibson (test) | Success Rate9 | 10 | |
| ObjectNav (Sketch goal) | Gibson tiny (test) | Success Rate6.5 | 10 | |
| ObjectNav (Sketch) | Gibson (test) | Success Rate6.5 | 10 | |
| RoomNav (Label) | Gibson (test) | Success Rate7.7 | 10 | |
| ViewNav (Edgemap goal) | Gibson tiny (test) | Success Rate60 | 10 | |
| ViewNav (Edgemap) | Gibson (test) | Success Rate60 | 10 |