Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ai 98% Match Research Paper AI Ethics Researchers,LLM Developers,Policy Makers,Social Scientists,HCI Researchers 1 week ago

Race and Gender in LLM-Generated Personas: A Large-Scale Audit of 41 Occupations

large-language-models › alignment
📄 Abstract

Abstract: Generative AI tools are increasingly used to create portrayals of people in occupations, raising concerns about how race and gender are represented. We conducted a large-scale audit of over 1.5 million occupational personas across 41 U.S. occupations, generated by four large language models with different AI safety commitments and countries of origin (U.S., China, France). Compared with Bureau of Labor Statistics data, we find two recurring patterns: systematic shifts, where some groups are consistently under- or overrepresented, and stereotype exaggeration, where existing demographic skews are amplified. On average, White (--31pp) and Black (--9pp) workers are underrepresented, while Hispanic (+17pp) and Asian (+12pp) workers are overrepresented. These distortions can be extreme: for example, across all four models, Housekeepers are portrayed as nearly 100\% Hispanic, while Black workers are erased from many occupations. For HCI, these findings show provider choice materially changes who is visible, motivating model-specific audits and accountable design practices.
Authors (7)
Ilona van der Linden
Sahana Kumar
Arnav Dixit
Aadi Sudan
Smruthi Danda
David C. Anastasiu
+1 more
Submitted
October 23, 2025
arXiv Category
cs.HC
arXiv PDF

Key Contributions

This large-scale audit of over 1.5 million occupational personas generated by four LLMs reveals systematic underrepresentation (White, Black) and overrepresentation (Hispanic, Asian) compared to BLS data, alongside stereotype exaggeration. It highlights significant racial and gender biases in LLM-generated portrayals across 41 occupations, demonstrating a critical need for improved fairness and alignment in generative AI.

Business Value

Crucial for developers and deployers of generative AI to understand and mitigate potential harms related to bias and fairness, ensuring responsible AI development and preventing reputational damage or discriminatory outcomes.