Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Rational Metareasoning for Large Language Models

Created by
  • Haebom

Author

C. Nicol o De Sabbata, Theodore R. Sumers, Badr AlKhamissi, Antoine Bosselut, Thomas L. Griffiths

Outline

This paper presents a novel approach to address the high cost of prompt-based inference methods that use additional computations in the inference process to improve the inference performance of large-scale language models (LLMs). Based on a computational model of meta-inference used in cognitive science, we propose a method to train LLMs to selectively use intermediate inference steps only when necessary. We develop a reward function that includes a penalty for unnecessary inferences, and train LLMs using it together with expert iteration. Experimental results show that the proposed method achieves 20-37% reduction in token generation for three models compared to the conventional few-shot chain-of-thought prompting and STaR, while maintaining task performance on a variety of datasets.

Takeaways, Limitations

Takeaways:
We present a novel method to effectively reduce the inference cost of LLM.
We demonstrate that the inference efficiency of LLM can be improved by utilizing the meta-inference model.
We demonstrate the effectiveness of cost savings without performance degradation across a variety of datasets.
Limitations:
Further studies are needed to investigate the generalization performance of the proposed reward function and its applicability to various LLMs.
Consideration should be given to the training costs and complexity of expert repetition methods.
Its effectiveness for specific types of reasoning problems needs to be verified, and further research is needed to determine whether it can be generalized to all types of problems.
👍