Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Mind to Hand: Purposeful Robotic Control via Embodied Reasoning

About

Humans act with context and intention, with reasoning playing a central role. While internet-scale data has enabled broad reasoning capabilities in AI systems, grounding these abilities in physical action remains a major challenge. We introduce Lumo-1, a generalist vision-language-action (VLA) model that unifies robot reasoning ("mind") with robot action ("hand"). Our approach builds upon the general multi-modal reasoning capabilities of pre-trained vision-language models (VLMs), progressively extending them to embodied reasoning and action prediction, and ultimately towards structured reasoning and reasoning-action alignment. This results in a three-stage pre-training pipeline: (1) Continued VLM pre-training on curated vision-language data to enhance embodied reasoning skills such as planning, spatial understanding, and trajectory prediction; (2) Co-training on cross-embodiment robot data alongside vision-language data; and (3) Action training with reasoning process on trajectories collected on Astribot S1, a bimanual mobile manipulator with human-like dexterity and agility. Finally, we integrate reinforcement learning to further refine reasoning-action consistency and close the loop between semantic inference and motor control. Extensive experiments demonstrate that Lumo-1 achieves significant performance improvements in embodied vision-language reasoning, a critical component for generalist robotic control. Real-world evaluations further show that Lumo-1 surpasses strong baselines across a wide range of challenging robotic tasks, with strong generalization to novel objects and environments, excelling particularly in long-horizon tasks and responding to human-natural instructions that require reasoning over strategy, concepts and space.

Peijun Tang, Shangjin Xie, Binyan Sun, Baifu Huang, Kuncheng Luo, Haotian Yang, Weiqi Jin, Jianan Wang• 2025

Related benchmarks

TaskDatasetResultRank
Spatial ReasoningEmbSpatial
Overall Accuracy75.6
30
Spatial ReasoningRefSpatial-Bench
Localization Score51.99
19
Object PlacementWhere2Place
Overall Score69.06
6
Spatial AwarenessSAT
Accuracy (All)74.67
6
Multi-modal ReasoningCV-Bench
Overall Accuracy86.36
6
Robotic Spatial ReasoningROBOSPATIAL
mAP (Mask)62.57
5
Showing 6 of 6 rows

Other info

Follow for update