Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

SASRA: Semantically-aware Spatio-temporal Reasoning Agent for Vision-and-Language Navigation in Continuous Environments

About

This paper presents a novel approach for the Vision-and-Language Navigation (VLN) task in continuous 3D environments, which requires an autonomous agent to follow natural language instructions in unseen environments. Existing end-to-end learning-based VLN methods struggle at this task as they focus mostly on utilizing raw visual observations and lack the semantic spatio-temporal reasoning capabilities which is crucial in generalizing to new environments. In this regard, we present a hybrid transformer-recurrence model which focuses on combining classical semantic mapping techniques with a learning-based method. Our method creates a temporal semantic memory by building a top-down local ego-centric semantic map and performs cross-modal grounding to align map and language modalities to enable effective learning of VLN policy. Empirical results in a photo-realistic long-horizon simulation environment show that the proposed approach outperforms a variety of state-of-the-art methods and baselines with over 22% relative improvement in SPL in prior unseen environments.

Muhammad Zubair Irshad, Niluthpol Chowdhury Mithun, Zachary Seymour, Han-Pang Chiu, Supun Samarasekera, Rakesh Kumar• 2021

Related benchmarks

TaskDatasetResultRank
Vision-Language NavigationR2R-CE (val-unseen)
Success Rate (SR)24
266
Vision-and-Language NavigationR2R-CE (val-seen)
SR36
49
Vision-and-Language NavigationVLN-CE 1.0 (val-seen)
Navigation Error (NE)7.17
20
Vision-and-Language NavigationVLN-CE 1.0 (val-unseen)
Navigation Error (NE)8.32
20
Vision-Language NavigationVLN-CE unseen (val)
NE (Navigation Error)8.32
8
Showing 5 of 5 rows

Other info

Follow for update