Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ai 95% Match Research Paper AI safety researchers,Robotics engineers,Formal methods experts,Developers of autonomous systems 3 weeks ago

SENTINEL: A Multi-Level Formal Framework for Safety Evaluation of LLM-based Embodied Agents

ai-safety › alignment
📄 Abstract

Abstract: We present Sentinel, the first framework for formally evaluating the physical safety of Large Language Model(LLM-based) embodied agents across the semantic, plan, and trajectory levels. Unlike prior methods that rely on heuristic rules or subjective LLM judgments, Sentinel grounds practical safety requirements in formal temporal logic (TL) semantics that can precisely specify state invariants, temporal dependencies, and timing constraints. It then employs a multi-level verification pipeline where (i) at the semantic level, intuitive natural language safety requirements are formalized into TL formulas and the LLM agent's understanding of these requirements is probed for alignment with the TL formulas; (ii) at the plan level, high-level action plans and subgoals generated by the LLM agent are verified against the TL formulas to detect unsafe plans before execution; and (iii) at the trajectory level, multiple execution trajectories are merged into a computation tree and efficiently verified against physically-detailed TL specifications for a final safety check. We apply Sentinel in VirtualHome and ALFRED, and formally evaluate multiple LLM-based embodied agents against diverse safety requirements. Our experiments show that by grounding physical safety in temporal logic and applying verification methods across multiple levels, Sentinel provides a rigorous foundation for systematically evaluating LLM-based embodied agents in physical environments, exposing safety violations overlooked by previous methods and offering insights into their failure modes.
Authors (13)
Simon Sinong Zhan
Yao Liu
Philip Wang
Zinan Wang
Qineng Wang
Zhian Ruan
+7 more
Submitted
October 14, 2025
arXiv Category
cs.AI
arXiv PDF

Key Contributions

Presents Sentinel, the first framework for formally evaluating the physical safety of LLM-based embodied agents across semantic, plan, and trajectory levels using formal temporal logic. It employs a multi-level verification pipeline to detect unsafe plans and trajectories before execution.

Business Value

Crucial for deploying embodied AI systems (e.g., robots) in safety-critical environments, ensuring reliability and public trust.