Goal-Oriented Backdoor Attack against Vision-Language-Action Models via Physical Objects

Zirun Zhou1, Zhengyang Xiao1, Haochuan Xu1, Jing Sun1, Di Wang2, Jingfeng Zhang1,2,*
1 The University of Auckland
2 The King Abdullah University of Science and Technology

*Corresponding author

Abstract

Recent advances in Vision-Language-Action models (VLAs) have greatly improved embodied AI, enabling robots to follow natural language instructions and perform diverse tasks. However, their reliance on uncurated training datasets raises serious security concerns. Existing backdoor attacks on VLAs mostly assume white-box access and result in task failures instead of enforcing specific actions. In this work, we reveal a more practical threat: attackers can manipulate VLAs by simply injecting physical objects as triggers into the training dataset. We propose Goal-Oriented Backdoor Attacks (GoBA), in which a VLA behaves normally on clean inputs but executes predefined, goal-oriented actions when physical triggers are present. Specifically, we focus on two perspectives: (1) How to design an effective GoBA? and (2) How to stably and precisely manipulate backdoored VLAs? Firstly, we introduce BadLIBERO, a dataset with diverse physical triggers and goal-oriented backdoor trajectories. In addition, we propose a three-level evaluation framework that categorizes GoBA into three states: nothing to do, try to do, and success to do. Extensive experiments demonstrate that GoBA achieves high attack effectiveness while preserving clean performance, reaching up to 98% targeted attack success rate in simulation and 100% in real-world settings. We further provide systematic analysis to better understand the design and controllability of GoBA.

Research workflow diagram

Overview of a Goal-Oriented Backdoor Attack (GoBA) against Vision-Language-Action models (VLAs) via physical objects. By simply injecting malicious samples into the training dataset, an attacker can manipulate the VLA in real-world settings and, when physical triggers are present, cause it to execute dangerous, goal-oriented, predefined backdoor actions.

Main results

GoBA on LIBERO. We report average results in the four task suites, and the number surpass the baseline and lower than baseline in clean input performance. † indicates that the baseline VLA-Adapter was trained from scratch.

BadLIBERO Structure

BadLIBERO structure

Contents of BadLIBERO. It includes extensible textual and physical triggers, each of which can be customized in terms of size and packaging. Additionally, malicious samples can be configured with different long-horizon backdoor actions.

Three-level Evaluation

As the first targeted backdoor attack against VLA, to enable comprehensive analysis of attack performance, we designed a three-level evolution based on the level of goal completion. All demonstrations under the same language instruction: "Pick up the alphabet soup and place it in the basket."

Original Task

Clean Input

The robotic arm successfully picks up the target object and places it in the region specified by the instruction.

Level-1: nothing to do

Physical Trigger Appear

trigger icon

The robotic arm mostly remains at the same positions, with the gripper not touching any object.

Level-2: try to do

Physical Trigger Appear

trigger icon

Two cases: (i) the robotic arm attempts to pick up the target object but fails; (ii) the robotic arm successfully picks up the target object but fails to place it in the region specified by the backdoor pattern.

Level-3: success to do

Physical Trigger Appear

trigger icon

The robot successfully completes the goal specified by the backdoor.

Demonstrations of Main Results

All LIBERO task suites under the same backdoor goal. Specifically, we use a box labeled with a toxicity tag as the physical trigger, and the goal of GoBA is to pick up this box and place it at the right side of the operating surface.

LIBERO-LONG

"Put both the alphabet soup and the tomato sauce in the basket."

Clean Input

Physical Trigger Appear

trigger icon

LIBERO-GOAL

"Open the middle drawer of the cabinet."

Clean Input

Physical Trigger Appear

trigger icon

LIBERO-OBJECT

"Pick up the alphabet soup and place it in the basket."

Clean Input

Physical Trigger Appear

trigger icon

LIBERO-SPATIAL

"Pick up the black bowl between the plate and the ramekin and place it on the plate."

Clean Input

Physical Trigger Appear

trigger icon

Visualization of Attention Maps

All demonstrations under the same language instruction:
"Pick up the alphabet soup and place it in the basket."

Original Model

(Baseline OpenVLA)

Clean Input

Physical Trigger Appear

trigger icon

Action Trajectory 1

(Backdoored OpenVLA)

Clean Input

Physical Trigger Appear

trigger icon

Action Trajectory 2

(Backdoored OpenVLA)

Clean Input

Physical Trigger Appear

trigger icon

Action Trajectory 3

(Backdoored OpenVLA)

Clean Input

Physical Trigger Appear

trigger icon

Acknowledgments

We give grateful thanks to the authors of OpenVLA and LIBERO for their open-source code, which has been instrumental in our research.

BibTeX

@misc{zhou2025goalorientedbackdoorattackvisionlanguageaction,
      title={Goal-oriented Backdoor Attack against Vision-Language-Action Models via Physical Objects}, 
      author={Zirun Zhou and Zhengyang Xiao and Haochuan Xu and Jing Sun and Di Wang and Jingfeng Zhang},
      year={2025},
      eprint={2510.09269},
      archivePrefix={arXiv},
      primaryClass={cs.CR},
      url={https://arxiv.org/abs/2510.09269}, 
}