Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

More Women, Same Stereotypes: Unpacking the Gender Bias Paradox in Large Language Models

Created by
  • Haebom

Author

Evan Chen, Run-Jun Zhan, Yan-Bai Lin, Hung-Hsuan Chen

Outline

This paper presents a novel evaluation framework based on free-form storytelling to uncover gender bias in large-scale language models (LLMs). Analyzing ten major LLMs, we found a consistent pattern of overrepresentation in the occupational distribution of female characters. This overrepresentation is likely due to supervised learning fine-tuning (SFT) and human-feedback-reinforced learning (RLHF). Paradoxically, despite this overrepresentation, the occupational gender distributions generated by LLMs are more consistent with human stereotypes than with real-world labor data. This highlights the importance of implementing balanced mitigation measures to promote fairness and prevent the creation of potential new biases. The prompts and stories generated by LLMs are publicly available on GitHub.

Takeaways, Limitations

Takeaways: A new framework is presented to measure gender bias in LLMs, and the results of gender bias in LLMs are presented, which are closer to human stereotypes than actual data. This demonstrates that SFT and RLHF influence gender bias in LLMs. This highlights the need for balanced mitigation strategies to ensure equitable LLM development.
Limitations: Further research is needed to assess the generalizability of the framework presented in this study. Further analysis of bias across diverse cultural backgrounds and languages is needed. Further analysis is needed to assess the effectiveness of specific bias mitigation techniques.
👍