Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Mechanistic Finetuning of Vision-Language-Action Models via Few-Shot Demonstrations

About

Vision-Language Action (VLAs) models promise to extend the remarkable success of vision-language models (VLMs) to robotics. Yet, unlike VLMs in the vision-language domain, VLAs for robotics require finetuning to contend with varying physical factors like robot embodiment, environment characteristics, and spatial relationships of each task. Existing fine-tuning methods lack specificity, adapting the same set of parameters regardless of a task's visual, linguistic, and physical characteristics. Inspired by functional specificity in neuroscience, we hypothesize that it is more effective to finetune sparse model representations specific to a given task. In this work, we introduce Robotic Steering, a finetuning approach grounded in mechanistic interpretability that leverages few-shot demonstrations to identify and selectively finetune task-specific attention heads aligned with the physical, visual, and linguistic requirements of robotic tasks. Through comprehensive on-robot evaluations with a Franka Emika robot arm, we demonstrate that Robotic Steering outperforms LoRA while achieving superior robustness under task variation, reduced computational cost, and enhanced interpretability for adapting VLAs to diverse robotic tasks.

Chancharik Mitra, Yusen Luo, Raj Saravanan, Dantong Niu, Anirudh Pai, Jesse Thomason, Trevor Darrell, Abrar Anwar, Deva Ramanan, Roei Herzig• 2025

Related benchmarks

TaskDatasetResultRank
Put MarkerReal-world Robotic Tasks
Success Rate80
7
Pick CubeReal on-robot tasks
Success Rate90
6
Place Cube in BowlReal on-robot tasks
Success Rate0.85
6
Push Cup to BowlReal on-robot tasks
Success Rate27.5
6
Press Button HardReal on-robot tasks
Success Rate8.50e+3
6
Robot ManipulationReal-world Robot Tasks Place Marker in Mug
Success Rate72.5
5
Robot ManipulationReal-world Robot Tasks Place Cube in Bowl
Success Rate65
5
Robot ManipulationReal-world Robot Tasks Pick Mug - Unseen
Success Rate65
5
Robot ManipulationReal-world Robot Tasks Place Marker in Mug - Lighting Variation
Success Rate0.475
5
Robot ManipulationReal-world Robot Tasks Place Marker in Mug - Form Variation
Success Rate37.5
5
Showing 10 of 11 rows

Other info

Follow for update