Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Generative AI tools are increasingly used to create portrayals of people in
occupations, raising concerns about how race and gender are represented. We
conducted a large-scale audit of over 1.5 million occupational personas across
41 U.S. occupations, generated by four large language models with different AI
safety commitments and countries of origin (U.S., China, France). Compared with
Bureau of Labor Statistics data, we find two recurring patterns: systematic
shifts, where some groups are consistently under- or overrepresented, and
stereotype exaggeration, where existing demographic skews are amplified. On
average, White (--31pp) and Black (--9pp) workers are underrepresented, while
Hispanic (+17pp) and Asian (+12pp) workers are overrepresented. These
distortions can be extreme: for example, across all four models, Housekeepers
are portrayed as nearly 100\% Hispanic, while Black workers are erased from
many occupations. For HCI, these findings show provider choice materially
changes who is visible, motivating model-specific audits and accountable design
practices.
Authors (7)
Ilona van der Linden
Sahana Kumar
Arnav Dixit
Aadi Sudan
Smruthi Danda
David C. Anastasiu
+1 more
Submitted
October 23, 2025
Key Contributions
This large-scale audit of over 1.5 million occupational personas generated by four LLMs reveals systematic underrepresentation (White, Black) and overrepresentation (Hispanic, Asian) compared to BLS data, alongside stereotype exaggeration. It highlights significant racial and gender biases in LLM-generated portrayals across 41 occupations, demonstrating a critical need for improved fairness and alignment in generative AI.
Business Value
Crucial for developers and deployers of generative AI to understand and mitigate potential harms related to bias and fairness, ensuring responsible AI development and preventing reputational damage or discriminatory outcomes.