Why Uncertainty Estimation Methods Fall Short in RAG An Axiomatic Analysis

Open Access
Authors
Publication date 2025
Host editors
  • Wanxiang Che
  • Joyce Nabende
  • Ekaterina Shutova
  • Mohammad Taher Pilehvar
Book title The 63rd Annual Meeting of the Association for Computational Linguistics (ACL 2025) : Findings of the Association for Computational Linguistics: ACL 2025
Book subtitle ACL 2025 : July 27-August 1, 2025
ISBN (electronic)
  • 9798891762565
Event 63rd Annual Meeting of the Association for Computational Linguistics, ACL 2025
Pages (from-to) 16596-16616
Number of pages 21
Publisher Kerrville, TX: Association for Computational Linguistics
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract

Large Language Models (LLMs) are valued for their strong performance across various tasks, but they also produce inaccurate or misleading outputs. Uncertainty Estimation (UE) quantifies the model's confidence and helps users assess response reliability. However, existing UE methods have not been thoroughly examined in scenarios like Retrieval-Augmented Generation (RAG), where the input prompt includes nonparametric knowledge. This paper shows that current UE methods cannot reliably estimate the correctness of LLM responses in the RAG setting. We propose an axiomatic framework to identify deficiencies in existing UE methods. Our framework introduces five constraints that an effective UE method should meet after incorporating retrieved documents into the LLM's prompt. Experimental results reveal that no existing UE method fully satisfies all the axioms, explaining their suboptimal performance in RAG. We further introduce a simple yet effective calibration function based on our framework, which not only satisfies more axioms than baseline methods but also improves the correlation between uncertainty estimates and correctness.

Document type Conference contribution
Language English
Published at https://doi.org/10.18653/v1/2025.findings-acl.852
Other links https://www.scopus.com/pages/publications/105028638773
Downloads
2025.findings-acl.852 (Final published version)
Permalink to this page
Back