TLDR: A new research paper introduces the “Quasi-Creature” and “Uncanny Valley of Agency” frameworks to explain user frustration with generative AI’s inconsistent performance. The “Move 78” experiment shows that AI’s failures, like lack of memory and generic outputs, lead to high user frustration, especially for experienced users. The paper advocates for AI design that transparently communicates limitations rather than aiming for perfect human simulation.
Generative Artificial Intelligence (AI) has brought about a fascinating paradox in our daily lives. These systems can produce incredibly fluent text, stunning images, and functional code, showcasing what seems like superhuman creativity and capacity. Yet, they often stumble with baffling, even absurd, failures in areas like common sense, consistency, and factual accuracy. This inconsistency isn’t just a technical glitch; it’s a fundamental challenge to how we understand and interact with these advanced technologies.
Introducing the Quasi-Creature
Researchers Mauricio Manhaes, Christine Miller, and Nicholas Schroeder propose that this profound frustration stems from the emergence of a new type of technological entity they call the “Quasi-Creature.” This is an agent that simulates intelligent behavior with remarkable sophistication but lacks the foundational elements of genuine understanding, such as embodiment and real-world interaction. When a Quasi-Creature fails, its errors aren’t seen as simple mechanical bugs, but rather as cognitive or intentional missteps, violating our deep-seated social expectations.
Navigating the Uncanny Valley of Agency
To explain the user experience with these Quasi-Creatures, the paper introduces a new conceptual space: the “Uncanny Valley of Agency.” This framework is distinct from Masahiro Mori’s original concept, which focused on physical appearance. Instead, the Uncanny Valley of Agency maps user trust and cognitive comfort against an entity’s perceived autonomous agency. The “valley” represents a sharp drop in user comfort that occurs when an entity appears highly capable and agentic but proves to be erratically and inscrutably unreliable. Its failures are perceived as breaches of understanding or intention, leading to significant cognitive dissonance.
Empirical Insights from the ‘Move 78’ Experiment
To ground this theoretical framework, the researchers conducted a mixed-methods study called “Move 78.” This experiment involved 37 participants, primarily students and faculty from art, design, and business management fields, who were already familiar with commercial generative AI tools. They collaborated with a customized generative AI system on a creative task, generating service concepts and keywords.
The AI system was intentionally designed to exhibit inconsistencies, mirroring the “unexplainable, unpredictable, and uncontrollable” behaviors often seen in current AI. Participants frequently reported frustration due to the AI’s lack of contextual memory, its tendency to produce generic or unhelpful outputs (like a recurring “Congratulations!” message for questions it couldn’t answer), and its misinterpretation of instructions.
Quantitative data from the NASA Task Load Index (NASA-TLX) strongly corroborated these qualitative observations. Participants reported exceptionally high levels of frustration (mean score of 15.08 out of 20) and mental demand (15.27). A key finding was an extremely strong negative correlation (r=−0.85) between perceived AI efficiency and user frustration: the more inefficient the AI was perceived to be, the higher the frustration levels soared.
Interestingly, the study also revealed an “Expert User Expectation Gap.” Participants with higher pre-experiment familiarity with AI reported significantly greater frustration. This suggests that experienced users, who likely have more sophisticated mental models of how a competent AI should “think,” experience a greater shock when the AI fails to meet these expectations, shattering their cognitive model of the AI’s “mind.”
The study also observed how social context influenced interaction. Dyadic groups (pairs) showed the most negative sentiment towards the AI but were the least likely to formally flag issues. This suggests that the presence of a partner allowed for shared commiseration, absorbing frustration internally rather than escalating it to the system.
Also Read:
- The Misunderstood Logic of AI: Why Humans Fail to Grasp AI’s Reasoning Steps
- Navigating the Social Landscape: How AI Agents Shape Our Identities and Interactions
Rethinking AI Design and Our Relationship with Technology
The findings from this research have significant implications for the future design, ethics, and societal integration of generative AI. The current approach of striving for seamless, flawless human simulation in AI design might be fundamentally flawed, as it inadvertently leads users into the Uncanny Valley of Agency by inflating expectations.
Instead, the authors suggest a more robust and ethical design philosophy: intentionally creating “seams” or “tells” in the user experience. These would serve as constant reminders of the AI’s alien nature and its statistical, non-semantic foundation, helping users manage their expectations and understand the system’s limitations. This approach aligns with the idea that technologies are not passive tools but active participants that shape human behavior and decision-making.
The paper concludes that understanding the Quasi-Creature and the Uncanny Valley of Agency is crucial for navigating our future with these powerful, yet alien, technologies responsibly. It calls for further research into how users adapt to these systems, how to design for “graceful failure,” and how to prevent the exploitation of user confusion for economic gain. To learn more about this groundbreaking research, you can read the full paper here.


