Dialect prejudice predicts AI decisions about people’s character, employability, and criminality
- Hundreds of millions of people now interact with language models, with uses ranging from serving as a writing aid to informing hiring decisions. Yet these language models are known to perpetuate systematic racial prejudices, making their judgments biased in problematic ways about groups like African Americans. While prior research has focused on overt racism in language models, social scientists have argued that racism with a more subtle character has developed over time. It is unknown whether this covert racism manifests in language models. Here, we demonstrate that language models embody covert racism in the form of dialect prejudice: we extend research showing that Americans hold raciolinguistic stereotypes about speakers of African American English and find that language models have the same prejudice, exhibiting covert stereotypes that are more negative than any human stereotypes about African Americans ever experimentally recorded, although closest to the ones from before the civil rights movement. By contrast, the language models’ overt stereotypes about African Americans are much more positive. We demonstrate that dialect prejudice has the potential for harmful consequences by asking language models to make hypothetical decisions about people, based only on how they speak. Language models are more likely to suggest that speakers of African American English be assigned less prestigious jobs, be convicted of crimes, and be sentenced to death. Finally, we show that existing methods for alleviating racial bias in language models such as human feedback training do not mitigate the dialect prejudice, but can exacerbate the discrepancy between covert and overt stereotypes, by teaching language models to superficially conceal the racism that they maintain on a deeper level. Our findings have far-reaching implications for the fair and safe employment of language technology.
SBIRs
- Finished and sent ONR email
- Worked on the white paper. Mostly collecting things and fleshing out the project.
- And I made a picture!

- 2:00 SimAccel meeting
- 3:05 LM collaboration meeting
- It’s interesting to me how these meetings went. Lots of discussion on how to integrate the work discussed in the white paper, but really, it was an excuse for them to “put AI in the system.” I think this is going to be hard to keep on track and the amount of money will pull everyone onto the project. And that will be the end of our IRAD department.
- 4:30 Book club
GPT-Agents
- 2:45 meeting. Will need to drop at 3:05. Made some organizational progress, and found out that there is no page limits, so the summaries don’t have to be so strict.
