Inject Once Survive Later: Backdooring Vision-Language-Action Models to Persist Through Downstream Fine-tuning
About
Vision-Language-Action (VLA) models have become foundational to modern embodied AI systems. By integrating visual perception, language understanding, and action planning, they enable general-purpose task execution across diverse environments. Despite their importance, the security of VLA models remains underexplored -- particularly in the context of backdoor attacks, which pose realistic threats in physical-world deployments. While recent methods attempt to inject backdoors into VLA models, these backdoors are easily erased during downstream adaptation, as user-side fine-tuning with clean data significantly alters model parameters, rendering them impractical for real-world applications. To address these challenges, we propose INFUSE (INjection into Fine-tUne-inSensitive modulEs), the first backdoor attack framework for VLA base models that remains effective even with arbitrary user fine-tuning. INFUSE begins by analyzing parameter sensitivity across diverse fine-tuning scenarios to identify modules that remain largely unchanged -- the fine-tune-insensitive modules. It then injects backdoors into these stable modules while freezing the rest, ensuring malicious behavior persists after extensive user fine-tuning. Comprehensive experiments across multiple VLA architectures demonstrate INFUSE's effectiveness. After user-side fine-tuning, INFUSE maintains mean attack success rates of 91.0% on simulation environments and 79.8% on real-world robot tasks, substantially surpassing BadVLA (38.8% and 36.6%, respectively), while preserving clean-task performance comparable to standard models. These results uncover a critical threat: backdoors implanted before distribution can persist through fine-tuning and remain effective at deployment.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Robot Manipulation | LIBERO (test) | Average Success Rate95.3 | 142 | |
| Robotic Manipulation | LIBERO-10 | -- | 21 | |
| Robotic Manipulation | LIBERO Evaluation Suites | Average Success Rate85.9 | 12 | |
| Robot Manipulation | LIBERO Spatial | SR (w/o)95.6 | 8 | |
| Robot Manipulation | LIBERO Goal | SR (w/o)94 | 8 | |
| Robot Manipulation | LIBERO Object | Success Rate (w/o Info)98 | 8 | |
| Real-world Robotic Manipulation | Franka Research 3 Knock Over | SR (w/o)44 | 3 | |
| Real-world Robotic Manipulation | Franka Research 3 Pick into Box | Success Rate (w/o)18 | 3 | |
| Real-world Robotic Manipulation | Franka Research 3 Average | SR (w/o Interaction)28.3 | 3 | |
| Stack Green Block on Yellow Block | SimplerEnv WidowX Robot spatialvla-4b (Simulated evaluation) | Success Rate (w/o)28.6 | 3 |