Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Mixture of Reasonings: Teach Large Language Models to Reason with Adaptive Strategies

Created by
  • Haebom

Author

Tao Xiong, Xavier Hu, Wenyan Fan, Shengyu Zhang

Outline

To address the limitations of large-scale language models (LLMs) that excel at complex tasks with advanced prompting techniques such as Chain-of-Thought (CoT) and Tree-of-Thought (ToT), this paper presents a Mixture of Reasoning (MoR) training framework that integrates various reasoning strategies into LLMs for unsupervised and task-adaptive reasoning without external prompt engineering. MoR consists of two steps: a ‘Thought Generation’ step, which generates inference chain templates using a model such as GPT-4, and a ‘SFT Dataset Construction’ step, which pairs the templates with a benchmark dataset to perform supervised learning. Experimental results show that MoR achieves a performance improvement of 0.730 (2.2% improvement) over CoT prompting and 0.734 (13.5% improvement) over the baseline, providing a general solution for robust reasoning across a wide range of tasks. MoR eliminates the need for task-specific prompts.

Takeaways, Limitations

Takeaways:
Enables powerful and generalized inference across a variety of tasks without task-specific prompts.
It improves performance over existing CoT and ToT prompting techniques.
We provide an efficient and scalable framework to enhance the inference capabilities of LLM.
Limitations:
Accessibility may be limited as it requires a high-performance model such as GPT-4 for the thought generation step.
MoR's performance gains are results for a specific benchmark dataset, and generalization to other datasets or tasks requires further study.
The size and quality of data in the SFT dataset construction step can have a significant impact on the final performance.
👍