TLDR: A new research paper investigates how to make robots explain their actions effectively. Through a user study, it found that people want explanations when robot behavior is surprising and prefer concise explanations that highlight the robot’s intention and relevant contextual factors. Based on these findings, the researchers developed algorithms for Belief-Desire-Intention (BDI) robots to detect surprising actions and generate clear, context-aware explanations, improving human-robot interaction.
Imagine a future where robots seamlessly assist us with daily chores, like tidying up the kitchen. While incredibly helpful, what happens when a robot does something unexpected, like opening the dishwasher before picking up cups from the table? This kind of surprising behavior can leave us confused, wondering why the robot acted that way. A new research paper explores how to make these interactions smoother by providing robots with the ability to explain themselves effectively.
The paper, titled “Effective Explanations for Belief-Desire-Intention Robots: When and What to Explain,” delves into the crucial aspects of robot explanations: when they should be given and what information they should contain. The goal is to prevent user annoyance while ensuring clarity and understanding.
Researchers Cong Wang, Roberto Calandra, and Verena Klös conducted a user study to understand human preferences for robot explanations. They found that people overwhelmingly want explanations when a robot’s actions are surprising. For instance, if a robot deviates from an expected sequence of actions, users are much more likely to desire an explanation. This highlights the importance of robots being able to detect when their behavior might be perceived as unusual by a human.
Beyond just when to explain, the study also shed light on what makes an explanation useful. Users preferred concise explanations that clearly stated the robot’s intention behind a confusing action and the specific contextual factors that influenced its decision. Explanations that were too broad (like just stating a high-level goal) or too detailed (listing every single belief or intention) were found to be less helpful. The most effective explanations focused on the “why” – the key reasons and conditions that led to the robot’s unexpected move.
Based on these valuable insights, the researchers developed two innovative algorithms. The first algorithm helps a robot identify when its actions might be surprising to a user. It does this by learning expected action sequences. If a robot performs an action that isn’t in its learned “expected successors” list, it triggers an explanation. This allows the robot to adapt and learn from interactions, becoming better at anticipating user confusion over time.
The second algorithm focuses on constructing the explanation itself. When an explanation is triggered, this algorithm gathers the relevant contextual information from the robot’s internal “Belief-Desire-Intention” (BDI) reasoning process. BDI is a common framework for designing intelligent agents, where robots act based on their beliefs about the world, their desires (goals), and their intentions (plans to achieve goals). The algorithm identifies the key beliefs and intentions that are most relevant to the surprising action, ensuring the explanation is focused and informative.
The researchers implemented these algorithms in a prototype system using JASON, a popular BDI-agent programming platform. This demonstration showed that their approach can be seamlessly integrated into existing robot control systems, making it practical for real-world applications. This work lays a strong foundation for creating robots that are not only capable but also transparent and understandable, leading to more natural and effective human-robot collaboration in our homes and beyond.
Also Read:
- Do AI Agents Practice What They Preach? Unpacking Belief-Behavior Consistency in LLM Simulations
- Decoding Algorithm Performance: How ‘Footprints’ Reveal Interactions with Problem Landscapes
For more in-depth information, you can read the full research paper here: Effective Explanations for Belief-Desire-Intention Robots: When and What to Explain.


