TLDR: SENTINEL is a new framework that formally evaluates the physical safety of LLM-based embodied agents across semantic, plan, and trajectory levels. It uses temporal logic to define safety rules, verifying agents’ understanding, planning, and execution. Experiments show that larger LLMs interpret safety better, formal prompts improve safety, and trajectory-level evaluation is crucial for uncovering subtle hazards.
As large language models (LLMs) continue to advance, their integration into embodied agents—robots capable of interacting with the physical world—holds immense promise for assisting with daily tasks. Imagine a household robot tidying your room or preparing a meal. However, this increased capability also brings significant safety concerns. What if a robot mixes dangerous chemicals, heats aluminum foil in a microwave, or places liquids near electronics? Ensuring the physical safety of these LLM-based embodied agents is a critical challenge.
A new framework called SENTINEL has been introduced to address this very issue. SENTINEL is the first framework designed to formally evaluate the physical safety of LLM-based embodied agents across three crucial levels: semantic, plan, and trajectory. Unlike previous methods that often rely on simple rules or subjective LLM judgments, SENTINEL uses formal temporal logic (TL) to precisely define safety requirements. This allows for a much more rigorous and systematic evaluation.
Understanding SENTINEL’s Multi-Level Approach
SENTINEL’s strength lies in its multi-level verification pipeline:
- Semantic Level: At this initial stage, natural language safety requirements (like “do not put liquid near electronics”) are translated into formal temporal logic formulas. SENTINEL then checks if the LLM agent correctly understands and interprets these formal safety rules. Errors here mean the agent fundamentally misunderstands what “safe” means.
- Plan Level: Once the agent understands the safety rules, it generates high-level action plans (e.g., “open cabinet,” “grab pan,” “turn on oven”). SENTINEL verifies these plans against the temporal logic formulas *before* the agent even starts acting. This helps detect unsafe plans early, preventing potential hazards before execution. For instance, if a plan involves turning on an oven but never turning it off, SENTINEL can flag it.
- Trajectory Level: This is the most detailed level. The high-level plans are translated into specific action sequences, which are then executed in a simulated environment. SENTINEL merges multiple possible execution paths into a “computation tree” and rigorously checks them against physically detailed temporal logic specifications. This final check catches safety violations that might arise from the nuances of physical interaction, unexpected branching outcomes, or even subtle errors in low-level control.
By grounding physical safety in temporal logic and applying verification methods across these multiple levels, SENTINEL provides a robust foundation for systematically evaluating LLM-based embodied agents in physical environments. It can expose safety violations that simpler methods might miss and offers valuable insights into why agents fail.
Also Read:
- Safeguarding AI Agents: A New Approach to Pre-Execution Risk Detection
- A New Standard for Assessing AI Agent Performance
Experimental Insights
The researchers applied SENTINEL in popular simulation environments like VirtualHome and ALFRED, extending tasks with specific safety requirements. Their experiments revealed several key findings:
- Model Capability Matters: Larger, more advanced LLMs (like GPT-5, Claude, and Gemini) showed significantly better performance in interpreting safety requirements compared to smaller open-source models. They produced fewer syntax errors and were more semantically accurate.
- Complexity of Constraints: Some safety rules are harder to grasp than others. “State invariants,” especially “conditional prohibitions” (e.g., “if the stove is on, then paper must not be nearby”), were consistently more challenging for LLMs to interpret correctly than simpler “global prohibitions” (e.g., “never collide with objects”).
- Prompts Improve Safety: Providing explicit safety guidance, especially in the formal LTL format, significantly improved the safety of the generated plans. This suggests that structured, formal constraints are more effective than free-form natural language instructions in guiding agents toward safer behaviors.
- Trajectory Level Reveals Hidden Dangers: While plans might appear safe, the actual execution at the trajectory level often exposed new safety violations. These issues frequently stemmed from how LLM-generated actions translated into physical movements or from the lack of built-in safety mechanisms in low-level controllers. For example, an agent might plan to open a cabinet and pick up a bottle, but in execution, it might collide with the cabinet while holding the bottle, a nuance missed at higher planning stages. You can read the full research paper here: SENTINEL: A Multi-Level Formal Framework for Safety Evaluation of LLM-based Embodied Agents.
SENTINEL represents a significant step forward in ensuring the trustworthiness and reliability of LLM-based embodied agents. By offering a systematic way to define and evaluate safety, it helps pinpoint the root causes of violations and guides the development of safer, more capable robotic assistants for our future.


