Teaching Parrots to See Red: Self-Audits of Generative Language Models Overlook Sociotechnical Harms
DOI:
https://doi.org/10.1609/aaaiss.v6i1.36070Abstract
The release of ChatGPT as a “low-key research preview” and its viral growth spurred a gold rush among tech companies marketing generative AI (GenAI) as a universal tool. In 2023, the U.S. secured voluntary commitments from top AI developers, including OpenAI, Google, Meta, and Anthropic, to conduct self-audits ensuring model safety before release. However, these models exhibit widespread biases, including by race and gender, unjustly discriminating against users. To inspect this contradiction, we review ten corporate self-audits, finding a notable absence of real-world use cases in sectors like education, creative works, and public policy. Instead, audits focus on thwarting adversarial consumers in hypothetical scenarios and rely on GenAI models to approximate human impacts. This approach places consumers at risk by impairing the mitigation of representational, allocational, and quality-of-service harms. We conclude with recommendations to address audit gaps and protect GenAI consumers.Downloads
Published
2025-08-01
How to Cite
Shieh, E., & Monroe-White, T. (2025). Teaching Parrots to See Red: Self-Audits of Generative Language Models Overlook Sociotechnical Harms. Proceedings of the AAAI Symposium Series, 6(1), 333–340. https://doi.org/10.1609/aaaiss.v6i1.36070
Issue
Section
Human-AI Collaboration: Exploring Diversity of Human Cognitive Abilities and Varied AI Models for Hybrid Intelligent Systems