Gated-Attention Architectures for Task-Oriented Language Grounding
About
To perform tasks specified by natural language instructions, autonomous agents need to extract semantically meaningful representations of language and map it to visual elements and actions in the environment. This problem is called task-oriented language grounding. We propose an end-to-end trainable neural architecture for task-oriented language grounding in 3D environments which assumes no prior linguistic or perceptual knowledge and requires only raw pixels from the environment and the natural language instruction as input. The proposed model combines the image and text representations using a Gated-Attention mechanism and learns a policy to execute the natural language instruction using standard reinforcement and imitation learning methods. We show the effectiveness of the proposed model on unseen instructions as well as unseen maps, both quantitatively and qualitatively. We also introduce a novel environment based on a 3D game engine to simulate the challenges of task-oriented language grounding over a rich set of instructions and environment states.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Vision-and-Language Navigation | Touchdown Seen (test) | TC11.9 | 13 | |
| Vision-and-Language Navigation | Touchdown Unseen (test) | nDTW4 | 11 | |
| Vision-and-Language Navigation | map2seq Seen (test) | nDTW30.1 | 10 | |
| Vision-and-Language Navigation | map2seq Unseen (test) | nDTW4.1 | 10 | |
| Outdoor Vision-and-Language Navigation | TOUCHDOWN (dev) | TC12.9 | 9 | |
| Outdoor Vision-and-Language Navigation | TOUCHDOWN (test) | Task Completion Rate (TC)13.9 | 9 | |
| Vision-and-Language Navigation | Touchdown seen (dev) | SDTW11.1 | 9 | |
| Vision-and-Language Navigation | map2seq unseen (dev) | nDTW3.9 | 8 | |
| Vision-and-Language Navigation | Touchdown unseen (dev) | SDTW1.3 | 7 | |
| Vision-and-Language Navigation | map2seq seen (dev) | SDTW17.2 | 7 |