TLDR: A study investigated how different types of AI explanations, particularly those showing uncertainty and global model behavior, influence human trust. Using the Unsupervised DeepView algorithm, researchers found that both numerical metrics and visual explanations successfully helped users calibrate their trust with the AI’s actual performance. However, the visual explanations did not lead to higher user satisfaction or direct trust compared to simpler numerical methods, possibly due to their complexity. The findings suggest that future Explainable AI (XAI) designs should prioritize clarity, offer tailored explanations for different user groups, and effectively combine local and global insights to improve user understanding and trust.
As artificial intelligence (AI) systems become more integrated into high-stakes fields like medical diagnosis and credit scoring, ensuring their trustworthiness is paramount. A key tool for building this trust is Explainable AI (XAI), which aims to make AI decisions understandable to humans. While much research focuses on XAI, a recent study delves into a less explored area: how explanations of AI uncertainty and global model behavior influence human trust.
The research, titled Uncertainty Awareness and Trust in Explainable AI – On Trust Calibration using Local and Global Explanations, was conducted by Carina Newen, Daniel Bodemer, Sonja Glantz, Emmanuel Müller, Magdalena Wischnewski, and Lenka Schnaubert. Their work highlights a critical gap: while many XAI methods exist, their effectiveness in calibrating human trust – aligning human trust with the AI’s actual trustworthiness – is often not rigorously evaluated, especially for explanations that cover the entire model’s behavior (global explanations) or its inherent uncertainties.
The study specifically investigated Unsupervised DeepView, an XAI algorithm designed to visualize uncertainties and robustness for complex, high-dimensional data. This algorithm was chosen because it offers a comprehensive view, including how robust the AI is against potential attacks and its overall certainty, making it potentially useful for both experts and non-experts.
To understand how different types of explanations affect trust, the researchers conducted an online study with nearly 200 participants. They designed two scenarios. In the first, participants evaluated AI models based solely on numerical metrics like test and validation accuracy. This served as a baseline to confirm that people could indeed distinguish between objectively better and worse models based on simple numbers. In the second scenario, a different group of participants was presented with visual explanations generated by Unsupervised DeepView, alongside numerical data. These visualizations offered a global overview of how many data points the AI considered ‘uncertain’ and provided specific examples of these uncertain instances.
The study aimed to test three main hypotheses: whether Unsupervised DeepView supports trust calibration, whether it leads to more satisfactory explanations than simple numerical methods, and whether it fosters more trust in the explanations themselves.
The findings revealed some interesting insights. Firstly, both numerical metrics and Unsupervised DeepView’s visual explanations successfully supported trust calibration. This means participants were able to correctly identify and trust the objectively more reliable AI models in both conditions. This is a crucial step, indicating that these explanation methods can help users align their trust with the AI’s actual performance.
However, when it came to user satisfaction and direct trust in the explanations, Unsupervised DeepView did not outperform simple numerical values. Participants did not find the visual explanations significantly more satisfying, and in some cases, they even mistrusted them more. The researchers suggest this might be due to the inherent complexity of a real-world explanation method like Unsupervised DeepView, which requires more effort to understand compared to straightforward percentages.
Through open-ended questions, participants provided valuable feedback. Many found the background estimations in Unsupervised DeepView confusing, suggesting that explanations should be as concise as possible, providing only the essential information. While some appreciated the visual representations, others highlighted the need for simpler color schemes and tailored explanations for different user groups (e.g., experts vs. laypersons). The study also found that users often combined both local (single instance examples) and global (overall model behavior) explanations to form their judgment of trustworthiness.
Also Read:
- Unveiling AI’s Inner Workings: Why Interpretability is Key to Trustworthy Systems
- Crafting Better Object Detection: A Visual Approach to Data Labeling
Based on these observations, the study derived several practical guidelines for designing future XAI algorithms. These include the importance of offering explanations tailored to the user’s background knowledge, simplifying visual elements like color schemes, and effectively combining both local and global explanations to enhance interpretability and trust calibration. This research underscores that while advanced visual explanations can help calibrate trust, their design must prioritize clarity and user-friendliness to be truly effective and foster user satisfaction.


