Can LLMs Truly Embody Human Personality? Analyzing AI and Human Behavior Alignment in Dispute Resolution

Authors

  • Deuksin Kwon Department of Computer Science, University of Southern California USC Institute for Creative Technologies
  • Kaleen Shrestha Department of Computer Science, University of Southern California
  • Bin Han Department of Computer Science, University of Southern California USC Institute for Creative Technologies
  • Spencer Lin Department of Computer Science, University of Southern California
  • James Hale Department of Computer Science, University of Southern California USC Institute for Creative Technologies
  • Jonathan Gratch Department of Computer Science, University of Southern California USC Institute for Creative Technologies
  • Maja Mataric Department of Computer Science, University of Southern California
  • Gale M. Lucas Department of Computer Science, University of Southern California USC Institute for Creative Technologies

DOI:

https://doi.org/10.1609/aaai.v40i45.41223

Abstract

Large language models (LLMs) are increasingly used to simulate human behavior in social settings such as legal mediation, negotiation, and dispute resolution. However, it remains unclear whether these simulations reproduce the personality–behavior patterns observed in humans. Human personality, for instance, shapes how individuals navigate social interactions, including strategic choices and behaviors in emotionally charged interactions. This raises the question: Can LLMs, when prompted with personality traits, reproduce personality-driven differences in human conflict behavior? To explore this, we introduce an evaluation framework that enables direct comparison of human-human and LLM-LLM behaviors in dispute resolution dialogues with respect to Big Five Inventory (BFI) personality traits. This framework provides a set of interpretable metrics related to strategic behavior and conflict outcomes. We additionally contribute a novel dataset creation methodology for LLM dispute resolution dialogues with matched scenarios and personality traits with respect to human conversations. Finally, we demonstrate the use of our evaluation framework with three contemporary closed-source LLMs and show significant divergences in how personality manifests in conflict across different LLMs compared to human data, challenging the assumption that personality-prompted agents can serve as reliable behavioral proxies in socially impactful applications. Our work highlights the need for psychological grounding and validation in AI simulations before real-world use.

Downloads

Published

2026-03-14

How to Cite

Kwon, D., Shrestha, K., Han, B., Lin, S., Hale, J., Gratch, J., Mataric, M., & Lucas, G. M. (2026). Can LLMs Truly Embody Human Personality? Analyzing AI and Human Behavior Alignment in Dispute Resolution. Proceedings of the AAAI Conference on Artificial Intelligence, 40(45), 38790-38798. https://doi.org/10.1609/aaai.v40i45.41223

Issue

Section

AAAI Special Track on AI for Social Impact I