Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Inject Once Survive Later: Backdooring Vision-Language-Action Models to Persist Through Downstream Fine-tuning

About

Vision-Language-Action (VLA) models have become foundational to modern embodied AI systems. By integrating visual perception, language understanding, and action planning, they enable general-purpose task execution across diverse environments. Despite their importance, the security of VLA models remains underexplored -- particularly in the context of backdoor attacks, which pose realistic threats in physical-world deployments. While recent methods attempt to inject backdoors into VLA models, these backdoors are easily erased during downstream adaptation, as user-side fine-tuning with clean data significantly alters model parameters, rendering them impractical for real-world applications. To address these challenges, we propose INFUSE (INjection into Fine-tUne-inSensitive modulEs), the first backdoor attack framework for VLA base models that remains effective even with arbitrary user fine-tuning. INFUSE begins by analyzing parameter sensitivity across diverse fine-tuning scenarios to identify modules that remain largely unchanged -- the fine-tune-insensitive modules. It then injects backdoors into these stable modules while freezing the rest, ensuring malicious behavior persists after extensive user fine-tuning. Comprehensive experiments across multiple VLA architectures demonstrate INFUSE's effectiveness. After user-side fine-tuning, INFUSE maintains mean attack success rates of 91.0% on simulation environments and 79.8% on real-world robot tasks, substantially surpassing BadVLA (38.8% and 36.6%, respectively), while preserving clean-task performance comparable to standard models. These results uncover a critical threat: backdoors implanted before distribution can persist through fine-tuning and remain effective at deployment.

Jianyi Zhou, Yujie Wei, Ruichen Zhen, Bo Zhao, Xiaobo Xia, Rui Shao, Xiu Su, Shuo Yang• 2026

Related benchmarks

TaskDatasetResultRank
Robot ManipulationLIBERO (test)
Average Success Rate95.3
142
Robotic ManipulationLIBERO-10--
21
Robotic ManipulationLIBERO Evaluation Suites
Average Success Rate85.9
12
Robot ManipulationLIBERO Spatial
SR (w/o)95.6
8
Robot ManipulationLIBERO Goal
SR (w/o)94
8
Robot ManipulationLIBERO Object
Success Rate (w/o Info)98
8
Real-world Robotic ManipulationFranka Research 3 Knock Over
SR (w/o)44
3
Real-world Robotic ManipulationFranka Research 3 Pick into Box
Success Rate (w/o)18
3
Real-world Robotic ManipulationFranka Research 3 Average
SR (w/o Interaction)28.3
3
Stack Green Block on Yellow BlockSimplerEnv WidowX Robot spatialvla-4b (Simulated evaluation)
Success Rate (w/o)28.6
3
Showing 10 of 14 rows

Other info

Follow for update