Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

LCLA: Language-Conditioned Latent Alignment for Vision-Language Navigation

About

We propose LCLA (Language-Conditioned Latent Alignment), a framework for vision-language navigation that learns modular perception-action interfaces by aligning sensory observations to a latent representation of an expert policy. The expert is first trained with privileged state information, inducing a latent space sufficient for control, after which its latent interface and action head are frozen. A lightweight adapter is then trained to map raw visual-language observations, via a frozen vision-language model, into the expert's latent space, reducing the problem of visuomotor learning to supervised latent alignment rather than end-to-end policy optimization. This decoupling enforces a stable contract between perception and control, enabling expert behavior to be reused across sensing modalities and environmental variations. We instantiate LCLA and evaluate it on a vision-language indoor navigation task, where aligned latent spaces yield strong in-distribution performance and robust zero-shot generalization to unseen environments, lighting conditions, and viewpoints while remaining lightweight at inference time.

Nitesh Subedi, Adam Haroon, Samuel Tetteh, Prajwal Koirala, Cody Fleming, Soumik Sarkar• 2026

Related benchmarks

TaskDatasetResultRank
Embodied NavigationRoom A (in-distribution)
Success Rate (SR)90.4
5
Embodied NavigationRoom B (out-of-distribution)
SR80.5
4
Showing 2 of 2 rows

Other info

Follow for update