Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Beyond Binary Rewards: Training LMs to Reason About Their Uncertainty

Created by
  • Haebom

Author

Mehul Damani, Isha Puri, Stewart Slocum, Idan Shenfeld, Leshem Choshen, Yoon Kim, Jacob Andreas

Outline

This paper demonstrates that training a language model (LM) to generate natural language inference processes via reinforcement learning (RL) improves performance on a variety of difficult question answering tasks. Existing RL methods use a binary reward function that only evaluates the accuracy of the LM output, resulting in side effects such as reduced calibration and increased generation of erroneous responses. In this paper, we propose Reinforcement Learning with Calibration Rewards (RLCR). RLCR is a method to simultaneously improve accuracy and calibrated confidence estimates by optimizing a reward function in which the LM generates both post-inference predictions and numerical confidence estimates, and adds Brier scores to the binary accuracy scores. We prove that this reward function yields models that generate accurate and well-calibrated predictions, and experimentally demonstrate that RLCR significantly improves calibration without compromising accuracy on a variety of datasets. We also show that the confidence expressed in language at test time can be used to improve accuracy and calibration through a confidence-weighted scaling method.

Takeaways, Limitations

Takeaways:
We demonstrate that reinforcement learning can improve the inference ability of language models while simultaneously improving confidence correction.
We show that a reward function using the Brier score is an effective way to improve both accuracy and power.
Presents a method to further improve performance by leveraging reliability information at test time.
We emphasize that the adjusted reliability estimation is important for creating more reliable inference models.
Limitations:
Further research is needed on the generalization performance of the proposed method.
Applicability verification for various types of inference problems is required.
Research into other correction score functions other than the Brier score may be needed.
👍