SELF-[IN]CORRECT: LLMs Struggle with Discriminating Self-Generated Responses

Authors

  • Dongwei Jiang Johns Hopkins University
  • Jingyu Zhang Johns Hopkins University
  • Orion Weller Johns Hopkins University
  • Nathaniel Weir Johns Hopkins University
  • Benjamin Van Durme Johns Hopkins University
  • Daniel Khashabi Johns Hopkins University

DOI:

https://doi.org/10.1609/aaai.v39i23.34603

Abstract

Can LLMs consistently improve their previous outputs for better results? For this to be true, LLMs would need to be better at discriminating among previously-generated alternatives, than generating initial responses. We explore the validity of this hypothesis in practice. We first formulate a unified framework that allows us to compare the generative and discriminative capability of any model on any task. In our resulting experimental analysis of several open-source and industrial LLMs, we observe that model’s are not reliably better at discriminating among previously-generated alternatives than generating initial responses. This finding challenges the notion that LLMs may be able to enhance their performance only through their own judgment.

Published

2025-04-11

How to Cite

Jiang, D., Zhang, J., Weller, O., Weir, N., Van Durme, B., & Khashabi, D. (2025). SELF-[IN]CORRECT: LLMs Struggle with Discriminating Self-Generated Responses. Proceedings of the AAAI Conference on Artificial Intelligence, 39(23), 24266–24275. https://doi.org/10.1609/aaai.v39i23.34603

Issue

Section

AAAI Technical Track on Natural Language Processing II