Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Reasoning about Uncertainty: Do Reasoning Models Know When They Don't Know?

Created by
  • Haebom

Author

Zhiting Mei, Christina Zhang, Tenny Yin, Justin Lidard, Ola Shorinwa, Anirudha Majumdar

Outline

This paper points out that although inference language models that enable multi-level inference via reinforcement learning have achieved state-of-the-art performance on several benchmarks, they are still vulnerable to hallucination phenomena, which confidently present incorrect answers. Therefore, it is important to quantify the confidence of the model in order to safely deploy the inference model in real-world applications. To this end, the paper explores uncertainty quantification of inference models and aims to answer three questions: whether the inference model is calibrated, how deeper inference affects model calibration, and whether models can improve calibration by explicitly inferring their inference process. Using introspective uncertainty quantification (UQ), we evaluate state-of-the-art inference models on various benchmarks. We find that inference models are generally overconfident, especially for incorrect answers, their self-expressed confidence estimates often exceed 85%, and that deeper inference exacerbates overconfidence, and that self-reflection can improve calibration (e.g., o3-Mini and DeepSeek R1), but not consistently across all models (e.g., Claude 3.7 Sonnet). Finally, we design essential UQ benchmarks and suggest important research directions to improve the calibration of inference models.

Takeaways, Limitations

Takeaways:
By revealing the problem of overconfidence in inference models and its severity, we emphasize the importance of uncertainty quantification for safe model deployment.
We present the possibility of improving the calibration of inference models through the introspective uncertainty quantification (UQ) technique.
We found that deeper reasoning can actually increase overconfidence.
Limitations:
Self-reflection techniques are not effective for all inference models.
Lack of UQ benchmarks.
Lack of specific methodology for improving the calibration of inference models.
👍