How Can Large Language Models Be More Reliable?

Authors

  • Yael Moros Daval Universitat Politècnica de València

DOI:

https://doi.org/10.1609/aies.v8i3.36790

Abstract

Large Language Models (LLMs) are increasingly integrated into everyday applications, yet their reliability remains inconsistent, even for seemingly simple queries. By “scaling up” and “shaping up”, these models have improved average accuracy and robustness to prompt variations, but they continue to display “difficulty discordance”: they solve harder tasks while making errors on easier ones. Moreover, they show a marked reluctance to refuse answers even when uncertain. Such behaviour deprives users of clear cues about when outputs can be trusted. This work explores strategies to enhance LLM reliability through confidence-based abstention, combining uncertainty estimation techniques with measures of question difficulty to define a model’s “safe operating area”. By ensuring that queries are either answered correctly or explicitly declined, the approach aims to enhance predictability, transparency, and user trust, while providing a framework for managing model limitations.

Downloads

Published

2025-10-15

How to Cite

Moros Daval, Y. (2025). How Can Large Language Models Be More Reliable?. Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society, 8(3), 2911-2912. https://doi.org/10.1609/aies.v8i3.36790