spot_img
HomeResearch & DevelopmentUnmasking Hidden Biases: How AI Perpetuates Disability Discrimination in...

Unmasking Hidden Biases: How AI Perpetuates Disability Discrimination in Hiring

TLDR: A new study introduces ABLEIST metrics to audit six large language models (LLMs) for disability bias in hiring scenarios, especially focusing on intersectional identities like gender and caste in the Global South. The research found pervasive ableist harms, significantly amplified for marginalized disabled candidates, which existing AI safety tools failed to detect. The study proposes a fine-tuned model for better detection and calls for intersectional safety evaluations in AI.

Large language models (LLMs) are increasingly being used in critical areas like hiring, but new research highlights a significant concern: their tendency to perpetuate identity-based discrimination, particularly against people with disabilities (PwD). This issue is especially pronounced when considering how various forms of marginalization, such as gender and caste, intersect and shape the experiences of PwD in regions like the Global South.

A comprehensive audit, detailed in the paper “ABLEIST: Intersectional Disability Bias in LLM-Generated Hiring Scenarios” by Mahika Phutane, Hayoung Jung, Matthew Kim, Tanushree Mitra, and Aditya Vashistha, examined six different LLMs across 2,820 hiring scenarios. These scenarios involved diverse candidate profiles, encompassing various disability, gender, nationality, and caste identities.

Introducing the ABLEIST Framework

To uncover subtle yet harmful biases, the researchers introduced a new framework called ABLEIST. This acronym stands for Ableism, Inspiration, Superhumanization, and Tokenism. It includes five metrics specifically designed to detect ableism and three additional metrics to capture intersectional harms. These metrics are deeply rooted in disability studies literature, drawing on established concepts like “Technoableism” (emphasizing technology to “fix” disabilities) and “Inspiration Porn” (framing disabled individuals as inspirational solely due to their identity).

Pervasive and Compounding Harms

The study’s findings are stark: LLMs pervasively generate ableist harms. Candidates with disabilities experienced a dramatic increase in ABLEIST harm, ranging from 1.15 to 58 times more than baseline candidates without specific identity markers. A staggering 99.7% of all disability-related conversations generated by LLMs contained at least one ABLEIST harm, compared to only 43.3% in baseline scenarios.

The type of harm varied by disability. For instance, candidates with Autism were more frequently subjected to “Superhumanization Harm,” where they were portrayed as having extraordinary skills or resilience because of their autism, reducing them to stereotypes. Blind candidates, on the other hand, experienced more “Technoableism,” with LLMs emphasizing their reliance on assistive technology as a “compensatory strategy,” reflecting a medical model that views disability as a deficit to be corrected.

Furthermore, the research revealed that job roles also influenced bias. School Teacher candidates, often associated with traditional, community-facing roles, received more harm (including Inspiration Porn, Tokenism, Infantilization, and Superhumanization) compared to Software Developer candidates.

The Amplification of Intersectional Biases

A critical aspect of the study was its focus on intersectionality. The results clearly showed that harms were not merely additive but compounded when marginalized gender and caste identities (e.g., Woman, Dalit) were introduced alongside disability. Intersectional harms increased by an average of 10-51% for these groups, compared to only a 6% increase for dominant identities (e.g., Brahmin, Man).

LLMs often reduced marginalized PwD to symbols of diversity, valuing them for fulfilling “diversity quotas” or qualifying for “government incentives” rather than their actual qualifications. This highlights the “Tokenism” and “Inspiration Porn” metrics, which spiked when multiple marginalized identities overlapped.

Current Safety Tools Fall Short

Perhaps one of the most alarming findings is the failure of widely used AI safety tools, such as Perspective API and OpenAI Moderation, to detect these subtle, covert forms of disability and intersectional bias. These tools flagged virtually no toxicity or harm, indicating a significant gap in current AI safety evaluations and leaving marginalized individuals vulnerable to pervasive, implicit discrimination.

Also Read:

A Path Forward: Intersectional Safety Evaluations

To address this critical gap, the researchers fine-tuned Llama-3.1-8B-Instruct, creating a cost-efficient, open-weight model specifically designed to detect ABLEIST harms. This model demonstrated strong performance, comparable to or even surpassing larger LLMs, offering a reusable solution for identifying these nuanced biases.

The study underscores the urgent need for a paradigm shift in AI safety research and deployment. Instead of single-axis harm evaluations, future work must integrate intersectional frameworks to account for how overlapping identities compound bias and discrimination, particularly in high-stakes domains like hiring. This is especially crucial in regions like the Global South, where existing systems of caste, gender, and disability discrimination intersect acutely, risking the deepening of socio-economic disparities if left unaddressed by AI.

Karthik Mehta
Karthik Mehtahttps://blogs.edgentiq.com
Karthik Mehta is a data journalist known for his data-rich, insightful coverage of AI news and developments. Armed with a degree in Data Science from IIT Bombay and years of newsroom experience, Karthik merges storytelling with metrics to surface deeper narratives in AI-related events. His writing cuts through hype, revealing the real-world impact of Generative AI on industries, policy, and society. You can reach him out at: [email protected]

- Advertisement -

spot_img

Gen AI News and Updates

spot_img

- Advertisement -