Determine whether elicited probabilities reflect LLMs’ true subjective beliefs
Determine whether probabilities elicited from large language models via natural-language prompting correspond to the models’ true subjective belief states—i.e., subjective probabilities that drive internal decision computations—or whether the elicited probabilities are superficial linguistic outputs that do not reflect the computations governing choices.
References
However, it is unclear whether stated probabilities reflect the model's 'true' beliefs: an elicited probability could track an internal epistemic state, or it could be a superficial linguistic output only weakly linked to the computations that drive choices (Pal et al., 2025; Wang et al., 2024a; Liu et al., 2024a).
— Do LLMs Act Like Rational Agents? Measuring Belief Coherence in Probabilistic Decision Making
(2602.06286 - Yamin et al., 6 Feb 2026) in Section 1 (Introduction)