What social stratifications in bias blind spot can tell us about implicit social bias in both LLMs and humans

Abstract Large language models (LLMs) are the engines behind generative Artificial Intelligence (AI) applications, the most well-known being chatbots. As conversational agents, they—much like the humans on whose data they are trained—exhibit social bias. The nature of social bias is that it unfairly...

Full description

Saved in:
Bibliographic Details
Main Authors: Sarah V. Bentley, David Evans, Claire K. Naughtin
Format: Article
Language:English
Published: Nature Portfolio 2025-08-01
Series:Scientific Reports
Online Access:https://doi.org/10.1038/s41598-025-14875-3
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1849226298238435328
author Sarah V. Bentley
David Evans
Claire K. Naughtin
author_facet Sarah V. Bentley
David Evans
Claire K. Naughtin
author_sort Sarah V. Bentley
collection DOAJ
description Abstract Large language models (LLMs) are the engines behind generative Artificial Intelligence (AI) applications, the most well-known being chatbots. As conversational agents, they—much like the humans on whose data they are trained—exhibit social bias. The nature of social bias is that it unfairly represents one group over another. Explicit bias has been observed in LLMs in the form of racial and gender bias, but we know little about less visible implicit forms of social bias. Highly prevalent in society, implicit human bias is hard to mitigate due to a well-replicated phenomenon known as introspection illusion. This describes a strong propensity to underestimate biases in oneself and overestimate them in others—a form of ‘bias blind spot’. In a series of studies with LLMs and humans, we measured variance in perception of others inherent in bias blind spot in order to determine if it was socially stratified, comparing ratings between LLMs and humans. We also assessed whether human bias blind spot extended to perceptions of LLMs. Results revealed perception of others to be socially stratified, such that those with lower educational, financial, and social status were perceived more negatively. This biased ‘perception’ was significantly more evident in the ratings of LLMs than of humans. Furthermore, we found evidence that humans were more blind to the likelihood of this bias in AI than in themselves. We discuss these findings in relation to AI, and how pursuing its responsible deployment requires contending with our own—as well as our technology’s—bias blind spots.
format Article
id doaj-art-b05b3996c49e4fbeab8ec799a40cd426
institution Kabale University
issn 2045-2322
language English
publishDate 2025-08-01
publisher Nature Portfolio
record_format Article
series Scientific Reports
spelling doaj-art-b05b3996c49e4fbeab8ec799a40cd4262025-08-24T11:29:04ZengNature PortfolioScientific Reports2045-23222025-08-0115111810.1038/s41598-025-14875-3What social stratifications in bias blind spot can tell us about implicit social bias in both LLMs and humansSarah V. Bentley0David Evans1Claire K. Naughtin2Data61, Commonwealth Scientific and Industrial Research Organisation (CSIRO)Data61, Commonwealth Scientific and Industrial Research Organisation (CSIRO)Data61, Commonwealth Scientific and Industrial Research Organisation (CSIRO)Abstract Large language models (LLMs) are the engines behind generative Artificial Intelligence (AI) applications, the most well-known being chatbots. As conversational agents, they—much like the humans on whose data they are trained—exhibit social bias. The nature of social bias is that it unfairly represents one group over another. Explicit bias has been observed in LLMs in the form of racial and gender bias, but we know little about less visible implicit forms of social bias. Highly prevalent in society, implicit human bias is hard to mitigate due to a well-replicated phenomenon known as introspection illusion. This describes a strong propensity to underestimate biases in oneself and overestimate them in others—a form of ‘bias blind spot’. In a series of studies with LLMs and humans, we measured variance in perception of others inherent in bias blind spot in order to determine if it was socially stratified, comparing ratings between LLMs and humans. We also assessed whether human bias blind spot extended to perceptions of LLMs. Results revealed perception of others to be socially stratified, such that those with lower educational, financial, and social status were perceived more negatively. This biased ‘perception’ was significantly more evident in the ratings of LLMs than of humans. Furthermore, we found evidence that humans were more blind to the likelihood of this bias in AI than in themselves. We discuss these findings in relation to AI, and how pursuing its responsible deployment requires contending with our own—as well as our technology’s—bias blind spots.https://doi.org/10.1038/s41598-025-14875-3
spellingShingle Sarah V. Bentley
David Evans
Claire K. Naughtin
What social stratifications in bias blind spot can tell us about implicit social bias in both LLMs and humans
Scientific Reports
title What social stratifications in bias blind spot can tell us about implicit social bias in both LLMs and humans
title_full What social stratifications in bias blind spot can tell us about implicit social bias in both LLMs and humans
title_fullStr What social stratifications in bias blind spot can tell us about implicit social bias in both LLMs and humans
title_full_unstemmed What social stratifications in bias blind spot can tell us about implicit social bias in both LLMs and humans
title_short What social stratifications in bias blind spot can tell us about implicit social bias in both LLMs and humans
title_sort what social stratifications in bias blind spot can tell us about implicit social bias in both llms and humans
url https://doi.org/10.1038/s41598-025-14875-3
work_keys_str_mv AT sarahvbentley whatsocialstratificationsinbiasblindspotcantellusaboutimplicitsocialbiasinbothllmsandhumans
AT davidevans whatsocialstratificationsinbiasblindspotcantellusaboutimplicitsocialbiasinbothllmsandhumans
AT claireknaughtin whatsocialstratificationsinbiasblindspotcantellusaboutimplicitsocialbiasinbothllmsandhumans