Estimating Semantic Alphabet Size for LLM Uncertainty Quantification

๐Ÿ“… 2025-09-17
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
Large language models (LLMs) suffer from biased semantic uncertainty estimation, poor interpretability, and hyperparameter sensitivity under few-shot settings. To address these issues, this paper proposes a hyperparameter-free semantic alphabet size correction method that leverages semantic clustering and resampling-based coverage analysis to accurately estimate discrete semantic entropyโ€”thereby mitigating underestimation caused by insufficient sampling. The approach preserves high interpretability while significantly improving uncertainty quantification accuracy in few-shot regimes. Empirically, it achieves performance on par with or superior to state-of-the-art methods in LLM hallucination detection. By enabling robust, transparent, and practical semantic uncertainty assessment, our method establishes a novel paradigm for black-box evaluation of LLMs.

Technology Category

Application Category

๐Ÿ“ Abstract
Many black-box techniques for quantifying the uncertainty of large language models (LLMs) rely on repeated LLM sampling, which can be computationally expensive. Therefore, practical applicability demands reliable estimation from few samples. Semantic entropy (SE) is a popular sample-based uncertainty estimator with a discrete formulation attractive for the black-box setting. Recent extensions of semantic entropy exhibit improved LLM hallucination detection, but do so with less interpretable methods that admit additional hyperparameters. For this reason, we revisit the canonical discrete semantic entropy estimator, finding that it underestimates the "true" semantic entropy, as expected from theory. We propose a modified semantic alphabet size estimator, and illustrate that using it to adjust discrete semantic entropy for sample coverage results in more accurate semantic entropy estimation in our setting of interest. Furthermore, our proposed alphabet size estimator flags incorrect LLM responses as well or better than recent top-performing approaches, with the added benefit of remaining highly interpretable.
Problem

Research questions and friction points this paper is trying to address.

Estimating semantic alphabet size for LLM uncertainty quantification
Improving semantic entropy estimation accuracy with sample coverage
Detecting LLM hallucinations while maintaining interpretability
Innovation

Methods, ideas, or system contributions that make the work stand out.

Modified semantic alphabet size estimator
Adjusts discrete semantic entropy for coverage
Improves accuracy while maintaining interpretability
๐Ÿ”Ž Similar Papers
No similar papers found.
L
Lucas H. McCabe
George Washington University
R
Rimon Melamed
George Washington University
T
Thomas Hartvigsen
University of Virginia
H. Howie Huang
H. Howie Huang
GraphLab, George Washington University
Graph AICyber SecurityComputer SystemsHigh-Performance Computing