Reasoning Models Better Express Their Confidence
Abstract
Despite their strengths, large language models (LLMs) often fail to communicate their confidence accurately, making it difficult to assess when they might be wrong and limiting their reliability. In this work, we demonstrate that reasoning models-LLMs that engage in extended chain-of-thought (CoT) reasoning-exhibit superior performance not only in problem-solving but also in accurately expressing their confidence. Specifically, we benchmark six reasoning models across six datasets and find that they achieve strictly better confidence calibration than their non-reasoning counterparts in 33 out of the 36 settings. Our detailed analysis reveals that these gains in calibration stem from the slow thinking behaviors of reasoning models-such as exploring alternative approaches and backtracking-which enable them to adjust their confidence dynamically throughout their CoT, making it progressively more accurate. In particular, we find that reasoning models become increasingly better calibrated as their CoT unfolds, a trend not observed in non-reasoning models. Moreover, removing slow thinking behaviors from the CoT leads to a significant drop in calibration. Lastly, we show that these gains are not exclusive to reasoning models-non-reasoning models also benefit when guided to perform slow thinking via in-context learning.
Community
🙁 LLMs are overconfident even when they are dead wrong.
🧐 What about reasoning models? Can they actually tell us “My answer is only 60% likely to be correct”?
❗Our paper suggests that they can! Through extensive analysis, we investigate what enables this emergent ability.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Do Reasoning Models Show Better Verbalized Calibration? (2025)
- Dynamic Early Exit in Reasoning Models (2025)
- Short-Path Prompting in LLMs: Analyzing Reasoning Instability and Solutions for Robust Performance (2025)
- ThinkEdit: Interpretable Weight Editing to Mitigate Overly Short Thinking in Reasoning Models (2025)
- Harnessing the Reasoning Economy: A Survey of Efficient Reasoning for Large Language Models (2025)
- Reasoning Models Know When They're Right: Probing Hidden States for Self-Verification (2025)
- When Thinking Fails: The Pitfalls of Reasoning for Instruction-Following in LLMs (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper