TLDR: This research paper explores the evolving roles of humans and AI in data science, proposing a collaborative model rather than replacement. Using the Truth, Beauty, and Justice (TBJ) framework, it highlights how human expertise remains crucial for ensuring accuracy, interpretability, and ethical application across the data science workflow’s planning, execution, and activation phases. The paper emphasizes that while AI excels at automating routine tasks, humans are indispensable for handling complex, uncertain, and ambiguous situations, mitigating biases, and translating insights into real-world impact. It also discusses the implications for the data science workforce, stressing the need for upskilling to focus on higher-order strategic and ethical considerations.
Artificial intelligence is rapidly changing the landscape of data science, automating tasks from survey creation to data analysis and summary writing. While this promises increased efficiency and quality, the reality is more nuanced. A recent research paper, AI, Humans, and Data Science: Optimizing Roles Across Workflows and the Workforce, explores how AI, humans, and data science can best work together, rather than AI simply replacing human roles.
The paper introduces a framework called Truth, Beauty, and Justice (TBJ) to evaluate AI, machine learning, and computational models for effective and ethical use. ‘Truth’ focuses on the accuracy, reliability, and validity of AI-generated plans, analyses, and insights. This means ensuring data is chosen correctly, synthetic data accurately reflects real-world patterns, and appropriate analytical methods are selected. Human oversight is crucial here to identify inaccuracies, biases, or inconsistencies that AI might miss.
‘Beauty’ goes beyond aesthetics to include the explainability of AI processes, the interpretability of their outputs, and the richness and depth of the insights generated. It emphasizes that valuable data analysis should uncover nuanced and comprehensive patterns, avoiding oversimplifications. Human data scientists are essential for contextualizing findings, identifying spurious correlations, and communicating complex results clearly.
‘Justice’ addresses the ethical implications of AI in data science, including data privacy, security, and the risk of biases in algorithms and data. It highlights the importance of identifying and mitigating biases to prevent unfair or discriminatory outcomes. The paper also touches on the impact of AI automation on the human workforce, particularly in entry-level data science roles, as an ethical consideration.
The paper distinguishes between different types of AI: Analytic AI (traditional machine learning for predictive analysis), Generative AI (creating new content like text or images), and AI Agents (systems that integrate multiple tools and can take autonomous action). AI agents are particularly relevant as they can automate complex workflows from data ingestion to visualization.
The data science workflow is divided into three phases: Planning, Execution, and Activation. In the Planning phase, which includes ideation and design, humans lead, complemented by generative AI. AI can help generate a diverse range of ideas, but human judgment is needed to distill and formalize these into accurate plans. The Execution phase, involving data gathering, processing, and analysis, sees AI agents taking a leading role in automating routine tasks. However, humans are vital for addressing ‘VUCA’ elements—Volatility, Uncertainty, Complexity, and Ambiguity—that AI struggles with. This includes understanding data dynamics, making trade-offs under uncertainty, disentangling complex interconnections, and clarifying ambiguous terms. Human data scientists ensure that analytical decisions are grounded in dynamic reality and that the results are fair and just. The paper warns against ‘blindness-by-design,’ where users might rely on AI without understanding the underlying assumptions or limitations, similar to how statistical software was sometimes used without full comprehension.
Finally, the Activation phase, where analytical results are turned into actionable insights, is human-led. While AI can synthesize information and identify patterns efficiently, human judgment is indispensable for interpreting findings within context, identifying potential biases or ‘hallucinations,’ and navigating complex ethical dilemmas. Humans are also better suited to understand organizational politics and change management for effective implementation of insights.
Also Read:
- Bridging the Gap: Visual Analytics for Transparent and Reliable AI
- AI’s Expanding Role in Scientific Discovery: From Assistants to Autonomous Researchers
The paper concludes that AI should strategically complement, not replace, human data scientists. This means a shift in the data science profession, with less demand for routine tasks and more emphasis on problem framing, strategic design, critical interpretation, and ethical considerations. It also highlights a potential challenge in the talent pipeline, as automation of junior roles could reduce opportunities for aspiring data scientists to gain foundational experience. Therefore, continuous training and upskilling are crucial for data science professionals to thrive in an AI-augmented world, focusing on higher-level skills, strategic thinking, and a deep understanding of ethical implications.


