Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Multimodal large language models (MLLMs) have advanced embodied agents by
enabling direct perception, reasoning, and planning task-oriented actions from
visual inputs. However, such vision driven embodied agents open a new attack
surface: visual backdoor attacks, where the agent behaves normally until a
visual trigger appears in the scene, then persistently executes an
attacker-specified multi-step policy. We introduce BEAT, the first framework to
inject such visual backdoors into MLLM-based embodied agents using objects in
the environments as triggers. Unlike textual triggers, object triggers exhibit
wide variation across viewpoints and lighting, making them difficult to implant
reliably. BEAT addresses this challenge by (1) constructing a training set that
spans diverse scenes, tasks, and trigger placements to expose agents to trigger
variability, and (2) introducing a two-stage training scheme that first applies
supervised fine-tuning (SFT) and then our novel Contrastive Trigger Learning
(CTL). CTL formulates trigger discrimination as preference learning between
trigger-present and trigger-free inputs, explicitly sharpening the decision
boundaries to ensure precise backdoor activation. Across various embodied agent
benchmarks and MLLMs, BEAT achieves attack success rates up to 80%, while
maintaining strong benign task performance, and generalizes reliably to
out-of-distribution trigger placements. Notably, compared to naive SFT, CTL
boosts backdoor activation accuracy up to 39% under limited backdoor data.
These findings expose a critical yet unexplored security risk in MLLM-based
embodied agents, underscoring the need for robust defenses before real-world
deployment.
Authors (10)
Qiusi Zhan
Hyeonjeong Ha
Rui Yang
Sirui Xu
Hanyang Chen
Liang-Yan Gui
+4 more
Submitted
October 31, 2025
Key Contributions
Introduces BEAT, the first framework to inject visual backdoors into MLLM-based embodied agents using objects as triggers. It addresses the challenge of trigger variability (viewpoints, lighting) by constructing a diverse training set and employing a novel two-stage training scheme (SFT + Contrastive Trigger Learning) to reliably implant these backdoors.
Business Value
Highlights critical security vulnerabilities in multimodal embodied AI systems, driving the development of more robust and secure AI agents for real-world applications like robotics and autonomous systems.