FLUE: Streamlined Uncertainty Estimation for Large Language Models

Authors

  • Shiqi Gao Beihang University
  • Tianxiang Gong Beihang University
  • Zijie Lin National University of Singapore
  • Runhua Xu Beihang University
  • Haoyi Zhou Beihang University
  • Jianxin Li Beihang University

DOI:

https://doi.org/10.1609/aaai.v39i16.33840

Abstract

Uncertainty estimation is essential for practical applications such as decision-making, risk assessment, and human-AI collaboration. However, Uncertainty estimation in open-ended question-answering (QA) tasks presents unique challenges. The output space for open-ended QA is vast and discrete, and the autoregressive nature of LLMs, combined with the rapid increase in model parameters, makes inference sampling significantly costly. An ideal uncertainty estimation for LLMs should meet two criteria: 1) incur no additional inference cost and 2) capture the semantic dependencies of token-level uncertainty within sequences. We propose a promising solution that converts redundancy into randomness in the extensive parameters of LLMs to quantify knowledge uncertainty. We can obtain token-level Monte Carlo samples without multiple inferences by introducing randomness during a single forward pass. We theoretically analyze the FLUE sampling method and employ a post-processing method to learn the state transitions from token uncertainty to sequence uncertainty. In open-ended question-answering tasks, we demonstrate that FLUE can achieve competitive performance in estimating the uncertainty of generated sentences without adding extra inference overhead.

Downloads

Published

2025-04-11

How to Cite

Gao, S., Gong, T., Lin, Z., Xu, R., Zhou, H., & Li, J. (2025). FLUE: Streamlined Uncertainty Estimation for Large Language Models. Proceedings of the AAAI Conference on Artificial Intelligence, 39(16), 16745-16753. https://doi.org/10.1609/aaai.v39i16.33840

Issue

Section

AAAI Technical Track on Machine Learning II