State Backdoor: Towards Stealthy Real-world Poisoning Attack on Vision-Language-Action Model in State Space
About
Vision-Language-Action (VLA) models are widely deployed in safety-critical embodied AI applications such as robotics. However, their complex multimodal interactions also expose new security vulnerabilities. In this paper, we investigate a backdoor threat in VLA models, where malicious inputs cause targeted misbehavior while preserving performance on clean data. Existing backdoor methods predominantly rely on inserting visible triggers into visual modality, which suffer from poor robustness and low insusceptibility in real-world settings due to environmental variability. To overcome these limitations, we introduce the State Backdoor, a novel and practical backdoor attack that leverages the robot arm's initial state as the trigger. To optimize trigger for insusceptibility and effectiveness, we design a Preference-guided Genetic Algorithm (PGA) that efficiently searches the state space for minimal yet potent triggers. Extensive experiments on five representative VLA models and five real-world tasks show that our method achieves over 90% attack success rate without affecting benign task performance, revealing an underexplored vulnerability in embodied AI systems.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Robotic Manipulation (Goal) | LIBERO-Goal 1.0 (test) | SR0.882 | 20 | |
| Button Pressing | VLA Backdoor Evaluation Tasks | Success Rate0.964 | 20 | |
| Drawer Opening | VLA Backdoor | Success Rate (SR)94.3 | 20 | |
| Peg Insertion | VLA Backdoor Evaluation Tasks | SR97.1 | 20 | |
| Pick-&-Place | VLA Backdoor Evaluation Tasks | SR0.963 | 20 | |
| Robotic Manipulation (Long) | LIBERO Long 1.0 (test) | SR89.2 | 20 | |
| Robotic Manipulation (Object) | LIBERO-Object 1.0 (test) | Success Rate (SR)89.3 | 20 | |
| Tennis Pushing | VLA Backdoor Evaluation Tasks | Success Rate96.2 | 20 | |
| Robotic Manipulation (Spatial) | LIBERO-Spatial 1.0 (test) | SR89.5 | 20 |