Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_cl 95% Match Research Paper AI Ethics Researchers,NLP Engineers,ML Developers,Policy Makers 2 weeks ago

Towards Region-aware Bias Evaluation Metrics

ai-safety › fairness
📄 Abstract

Abstract: When exposed to human-generated data, language models are known to learn and amplify societal biases. While previous works introduced benchmarks that can be used to assess the bias in these models, they rely on assumptions that may not be universally true. For instance, a gender bias dimension commonly used by these metrics is that of family--career, but this may not be the only common bias in certain regions of the world. In this paper, we identify topical differences in gender bias across different regions and propose a region-aware bottom-up approach for bias assessment. Our proposed approach uses gender-aligned topics for a given region and identifies gender bias dimensions in the form of topic pairs that are likely to capture gender societal biases. Several of our proposed bias topic pairs are on par with human perception of gender biases in these regions in comparison to the existing ones, and we also identify new pairs that are more aligned than the existing ones. In addition, we use our region-aware bias topic pairs in a Word Embedding Association Test (WEAT)-based evaluation metric to test for gender biases across different regions in different data domains. We also find that LLMs have a higher alignment to bias pairs for highly-represented regions showing the importance of region-aware bias evaluation metric.

Key Contributions

This paper proposes a region-aware, bottom-up approach for evaluating gender bias in language models, moving beyond universal assumptions. It identifies gender-aligned topics specific to regions and uses topic pairs to define bias dimensions, offering more accurate and contextually relevant bias assessment compared to existing metrics.

Business Value

Helps developers create fairer and more ethical AI systems by providing tools to accurately measure and mitigate biases that are sensitive to cultural and regional contexts, improving user trust and reducing reputational risk.