TLDR: A new study introduces a method to understand users’ ‘soft ethics’ (moral preferences) for software development. It uses immersive role-playing games to collect rich, contextual data on decision-making in digital privacy scenarios. This qualitative data is then interpreted by a customized LLM (‘GPT Anthropologist’) to create individual ethical profiles. The research found that this narrative-driven, anthropologically-informed approach significantly improves the LLM’s ability to predict user behavior, paving the way for more human-aligned AI systems.
In an increasingly digital world, where autonomous systems and software play a significant role in our daily lives, ensuring these technologies align with human values is paramount. Traditional methods for understanding user ethics often fall short, struggling to capture the nuanced, context-dependent nature of our moral decisions. A recent study introduces a groundbreaking approach that combines immersive role-playing games with advanced AI analysis to better understand these complex human moral preferences.
The research, titled An Anthropologist LLM to Elicit Users’ Moral Preferences Through Role-Play, by Gianluca De Ninno, Paola Inverardi, and Francesca Belotti, delves into what they term “soft ethics.” Unlike “hard ethics” which shapes laws and regulations, soft ethics refers to the individual moral preferences that guide our behavior within the vast space of decisions not dictated by law. Capturing these subtle preferences is crucial for developing software that truly resonates with users.
A Novel Blend of Anthropology and AI
The core of this innovative method lies in its interdisciplinary approach, drawing heavily from narrative and phenomenological anthropology. Instead of trying to categorize values into predefined universal sets, the study assumes that moral values are dynamic, shaped by lived experiences, cultural backgrounds, and social interactions. This perspective allows for a richer, more situated understanding of how ethical considerations emerge in real-time.
To achieve this, participants engaged in immersive role-playing games (RPGs) designed around ethically charged scenarios, specifically in the domain of digital privacy. RPGs were chosen for their unique ability to simulate complex social and contextual dynamics, allowing researchers to observe moral decision-making in situations that closely resemble real life. During these sessions, participants created characters, made choices, and interacted with non-player characters (NPCs) and even a customized GPT-4o model, referred to as the “GPT Anthropologist,” which acted as an “oracle” for guidance.
Gathering Rich, Contextual Data
The data collection process was multifaceted, capturing a comprehensive picture of each participant’s moral reasoning. It included:
-
Self-Descriptions: Adjectives chosen by players to describe their personality and social attitudes.
-
Game Diaries: Participants documented their reasoning, emotional responses, conflicts of values, and how group dynamics influenced their decisions.
-
GPT Prompts: Records of questions participants asked the “GPT oracle” during the game, revealing their digital literacy and areas of uncertainty.
-
Researcher’s Fieldnotes: Observations on players’ behavioral patterns, decisions, and interactions.
-
User Stories: Personal experiences shared during follow-up focus groups, bridging in-game decisions with real-world concerns.
This wealth of qualitative data was then organized into individual “narrative tables” for each participant. These tables provided the LLM with comprehensive access to contextual, behavioral, social, and experiential factors.
The GPT Anthropologist: Interpreting Human Morality
The customized GPT-4o model, “GPT Anthropologist,” was trained with anthropological frameworks and glossaries, along with detailed information about the game context. Its task was to analyze the co-constructed narrative data through the lens of lived experience and relational reasoning, generating individual ethical profiles for each player. These profiles went beyond simple decisions, considering complex interactions, mutual influences, and implicit values to identify underlying moral patterns.
The study’s evaluation showed significant improvements in the model’s predictive accuracy. When provided with the rich narrative tables and guided by an anthropological framework, GPT-A achieved an 80% accuracy in predicting user responses to new ethical scenarios, outperforming less informed configurations of the LLM. This highlights that understanding the context, motivations, relational dynamics, and past experiences behind a decision provides meaningful insight into behavioral patterns, enabling accurate predictions of future decisions.
Also Read:
- Evaluating LLMs for Automated Ethical Reasoning in Software Engineering
- Unpacking AI’s Social Compass: How Language Models Navigate Role Conflicts and Reveal Hidden Biases
Towards More Human-Aligned AI
The findings suggest that LLMs can be effectively employed to automate and enhance the understanding of user moral preferences and decision-making processes, particularly in the early stages of software development. By integrating contextual knowledge and an interpretive lens into AI, this approach not only enhances AI explainability but also ensures a human-centric perspective in requirement elicitation.
This research lays the groundwork for AI assistants capable of recognizing and adapting to individuals’ soft ethics, moving beyond surface-level processing to engage in deeper contextual interpretation. While future work aims to expand the interdisciplinary scope and automate aspects of the game master role, this foundational study demonstrates a powerful new way to align AI systems more closely with the complex and fluid nature of human values.


