Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Ensemble Learning for Large Language Models in Text and Code Generation: A Survey

Created by
  • Haebom

Author

Mari Ashiga, Wei Jie, Fan Wu, Vardan Voskanyan, Fateme Dinmohammadi, Paul Brookes, Jingzhi Gong, Zheng Wang

Outline

This paper examines ensembling techniques for large-scale language models (LLMs) based on generative pre-trained transformers (GPTs). Individual LLMs often produce inconsistent outputs and exhibit bias, limiting their ability to adequately represent diverse linguistic patterns. Furthermore, many powerful LLMs are closed-source, limiting their industrial applications due to data privacy concerns. Building on their success in text generation, this paper examines LLM ensemble techniques for code generation and analyzes their capabilities by categorizing them into seven key approaches: weighted merging, knowledge fusion, expert mixing, reward ensemble, output ensemble, routing, and cascading. We highlight key advantages, including enhanced representation of diversity, improved output quality, and increased application flexibility. This approach aids in model selection for practical tasks and lays the foundation for extending ensemble strategies to multimodal LLMs.

Takeaways, Limitations

Takeaways:
LLM ensemble techniques offer the potential to enhance diversity representation, improve output quality, and increase application flexibility.
Providing effective model selection criteria through an analysis of the characteristics, pros and cons of seven major LLM ensemble methods.
Suggesting the possibility of extending ensemble strategies to multimodal LLMs.
Limitations:
This paper focuses on reviewing existing research and does not include proposals for new ensemble techniques or experimental results.
The lack of performance comparison and analysis of each ensemble technique may result in a lack of clear guidance on selecting the optimal technique for practical application.
Lack of specific suggestions for extending ensemble strategies to multimodal LLM.
👍