Fairness Perceptions of Large Language Models
DOI:
https://doi.org/10.1609/aaai.v40i42.40848Abstract
Large language models (LLMs) are increasingly used for decision-making tasks where fairness is an essential desideratum. But what does fairness even mean to an LLM? To investigate this, we conduct a comprehensive evaluation of how LLMs perceive fairness in the context of resource allocation, using both synthetic and real-world data. We find that several state-of-the-art LLMs, when instructed to be fair, tend to prioritize improving collective welfare rather than distributing benefits equally. Their perception of fairness is somewhat sensitive to how user preferences are represented, but less so to the real-world context of the decision-making task. Finally, we show that the best strategy for aligning an LLM's perception of fairness to a specific criterion is to provide it as a mathematical objective, without referencing "fairness", as this prevents the LLM from mixing the criterion with its own prior notions of fairness. Our results provide practical insights for understanding and shaping how LLMs interpret fairness in resource allocation problems.Downloads
Published
2026-03-14
How to Cite
Cookson, B., Ebadian, S., & Shah, N. (2026). Fairness Perceptions of Large Language Models. Proceedings of the AAAI Conference on Artificial Intelligence, 40(42), 35393–35401. https://doi.org/10.1609/aaai.v40i42.40848
Issue
Section
AAAI Technical Track on Philosophy and Ethics of AI