Improving LLM Group Fairness on Tabular Data via In-Context Learning

Authors

  • Valeriia Cherepanova Amazon AWS AI
  • Chia-Jung Lee Amazon AWS AI
  • Nil-Jana Akpinar Amazon AWS AI
  • Riccardo Fogliato Amazon AWS AI
  • Martin Bertran Lopez Amazon AWS AI
  • Michael Kearns Amazon AWS AI University of Pennsylvania
  • James Zou Amazon AWS AI Stanford University

DOI:

https://doi.org/10.1609/aies.v8i1.36572

Abstract

Large language models (LLMs) have been shown to be effective on tabular prediction tasks in the low-data regime, leveraging their internal knowledge and ability to learn from instructions and examples. However, LLMs can fail to generate predictions that satisfy group fairness, that is, produce equitable outcomes across groups. Critically, conventional debiasing approaches for natural language tasks do not directly translate to mitigating group unfairness in tabular settings. In this work, we systematically investigate four empirical approaches to improve group fairness of LLM predictions on tabular datasets, including fair prompt optimization, soft prompt tuning, strategic selection of few-shot examples, and self-refining predictions via chain-of-thought reasoning. Through experiments on four tabular datasets using both open-source and proprietary LLMs, we show the effectiveness of these methods in enhancing demographic parity while maintaining high overall performance. Our analysis provides actionable insights for practitioners in selecting the most suitable approach based on their specific requirements and constraints.

Downloads

Published

2025-10-15

How to Cite

Cherepanova, V., Lee, C.-J., Akpinar, N.-J., Fogliato, R., Bertran Lopez, M., Kearns, M., & Zou, J. (2025). Improving LLM Group Fairness on Tabular Data via In-Context Learning. Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society, 8(1), 579-590. https://doi.org/10.1609/aies.v8i1.36572