[공지사항]을 빙자한 안부와 근황 
Show more

Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Reasoning about Uncertainty: Do Reasoning Models Know When They Don't Know?

Created by
  • Haebom

Author

Zhiting Mei, Christina Zhang, Tenny Yin, Justin Lidard, Ola Shorinwa, Anirudha Majumdar

Outline

This paper focuses on the fact that inference language models that enable multi-level inference via reinforcement learning have achieved state-of-the-art performance on many benchmarks, but, like traditional language models, suffer from hallucinations, which confidently give incorrect answers. In order to safely deploy inference models in real-world applications, it is important to understand the reliability of the model. Therefore, in this paper, we explore uncertainty quantification of inference models and answer three questions: whether inference models are calibrated, how deeper inference affects model calibration, and whether calibration can be improved by explicitly inferring the inference process. To this end, we introduce introspective uncertainty quantification (UQ) and evaluate state-of-the-art inference models on various benchmarks. Our experimental results show that inference models are generally overconfident, especially when the self-verbalized confidence estimates for incorrect answers often exceed 85%, and that while deeper inference exacerbates overconfidence, introspection can sometimes improve calibration. Finally, we design essential UQ benchmarks and suggest important research directions for improving the calibration of inference models.

Takeaways, Limitations

Takeaways:
Revealing the problem of overconfidence in inference models and its severity.
Suggesting the possibility of improving model calibration through introspective uncertainty quantification (introspective UQ).
Emphasizes the importance of developing UQ benchmarks to improve the reliability of inference models.
Limitations:
Corrective improvements through self-reflection do not apply to all models (lack of consistency).
The need to develop more comprehensive and rigorous UQ benchmarks.
👍