Bias Unveiled: Investigating Social Bias in LLM-Generated Code

Authors

  • Lin Ling Concordia University
  • Fazle Rabbi Concordia University
  • Song Wang York University
  • Jinqiu Yang Concordia University

DOI:

https://doi.org/10.1609/aaai.v39i26.34961

Abstract

Large language models (LLMs) have significantly advanced the field of automated code generation. However, a notable research gap exists in evaluating social biases that may be present in the code produced by LLMs. To solve this issue, we propose a novel fairness framework, i.e., Solar, to assess and mitigate the social biases of LLM-generated code. Specifically, Solar can automatically generate test cases for quantitatively uncovering social biases of the auto-generated code by LLMs. To quantify the severity of social biases in generated code, we develop a dataset that covers a diverse set of social problems. We applied Solar and the crafted dataset to four state-of-the-art LLMs for code generation. Our evaluation reveals severe bias in the LLM-generated code from all the subject LLMs. Furthermore, we explore several prompting strategies for mitigating bias, including Chain-of-Thought (CoT) prompting, combining positive role-playing with CoT prompting and dialogue with Solar. Our experiments show that dialogue with Solar can effectively reduce social bias in LLM-generated code by up to 90%. Last, we make the code and data publicly available is highly extensible to evaluate new social problems.

Published

2025-04-11

How to Cite

Ling, L., Rabbi, F., Wang, S., & Yang, J. (2025). Bias Unveiled: Investigating Social Bias in LLM-Generated Code. Proceedings of the AAAI Conference on Artificial Intelligence, 39(26), 27491–27499. https://doi.org/10.1609/aaai.v39i26.34961

Issue

Section

AAAI Technical Track on AI Alignment