Notebookcheck Logo

New research exposes AI’s lingering bias against African American English dialects

"Matched guise probing” was used by researchers to uncover prejudiced biases. (Image source: Dall-E 3)
"Matched guise probing” was used by researchers to uncover prejudiced biases. (Image source: Dall-E 3)
Recent research uncovers covert biases in AI language models, particularly against African American English (AAE). These models continue to associate AAE with negative stereotypes, possibly influencing future decisions in employment and criminal justice. The study uses "matched guise probing" for proof of concept.

A new study has exposed the covert racism embedded within AI language models, particularly in their treatment of African American English (AAE). Unlike previous research focusing on overt racism (like the CrowS-Pairs study to measure social biases in Masked LLMs), this study places special emphasis on how AI models subtly perpetuate negative stereotypes through dialect prejudice. These biases are not immediately visible but manifest obviously, such as associating AAE speakers with lower-status jobs and harsher criminal judgments.

The study found that even models trained to reduce overt bias still harbor deep-seated prejudices. This could have far-reaching implications, especially as AI systems become increasingly integrated into critical areas like employment and criminal justice, where fairness and equity is critical above all else.

Green text is Standard American English, while blue is African American English. In figure 'd', predictions for the SAE and AAE inputs are illustrated by five adjectives. (Image source: Nature)
Green text is Standard American English, while blue is African American English. In figure 'd', predictions for the SAE and AAE inputs are illustrated by five adjectives. (Image source: Nature)

The researchers employed a technique called “matched guise probing” to uncover these biases. By comparing how AI models responded to texts written in Standard American English (SAE) versus AAE, they were able to demonstrate that the models consistently associate AAE with negative stereotypes, even when the content was identical. This is a clear indicator of a fatal flaw in current AI training methods — surface-level improvements in reducing overt racism do not necessarily translate to the elimination of deeper, more insidious forms of bias.

AI will undoubtedly continue to evolve and integrate into more aspects of society. However, that also raises the risk of perpetuating and even amplifying existing societal inequalities, rather than mitigating them. Scenarios like these are the reason these discrepancies should be addressed as a priority.

Figure 'a' shows language modelling perplexity and stereotype strength on AAE text as a function of model size, while figure 'b' signifies change in stereotype strength and favourability. (Image source: Nature)
Figure 'a' shows language modelling perplexity and stereotype strength on AAE text as a function of model size, while figure 'b' signifies change in stereotype strength and favourability. (Image source: Nature)
Strongest stereotypes about African Americans in humans vs overt and covert stereotypes in Language models. (Image source: Nature)
Strongest stereotypes about African Americans in humans vs overt and covert stereotypes in Language models. (Image source: Nature)

Source(s)

Read all 8 comments / answer
static version load dynamic
Loading Comments
Comment on this article
Please share our article, every link counts!
Mail Logo
> Expert Reviews and News on Laptops, Smartphones and Tech Innovations > News > News Archive > Newsarchive 2024 08 > New research exposes AI’s lingering bias against African American English dialects
Anubhav Sharma, 2024-08-29 (Update: 2024-08-29)