Separating Aleatoric and Epistemic Uncertainty in Single-Model Semantic Entropy
Develop a single-model uncertainty quantification method for large language models that separates aleatoric uncertainty from epistemic uncertainty when computing semantic entropy over bidirectional-entailment semantic clusters, ensuring that the entropy measure is not artificially inflated by hallucinated, diverse but factually incorrect outputs due to lack of knowledge.
References
First, intra-model semantic entropy may conflate aleatoric and epistemic sources at the single-model level: an LLM that lacks knowledge about a prompt may hallucinate diverse but factually incorrect answers across samples, artificially inflating SE beyond pure aleatoric uncertainty~\citep{shorinwa2025survey}. This epistemic contamination of SE is a known open problem in LLM uncertainty quantification and is orthogonal to the multi-LLM contribution of CoE.