Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Prompting is not Enough: Exploring Knowledge Integration and Controllable Generation on Large Language Models

Created by
  • Haebom

Author

Tingjia Shen, Hao Wang, Chuan Qin, Ruijun Sun, Yang Song, Defu Lian, Hengshu Zhu, Enhong Chen

Outline

This paper proposes the GenKI framework, which aims to improve the performance of open-ended question answering (OpenQA) using large-scale language models (LLMs). GenKI simultaneously explores knowledge integration and controllable generation, addressing two key challenges of LLM-based OpenQA methods: effective knowledge integration and adaptive generation of specific answer formats for diverse task situations. To achieve this, we introduce a knowledge integration model that retrieves relevant knowledge, fine-tunes the retrieved knowledge, and enables controllable generation through ensemble models based on consistency, fluency, and answer format guarantees. Experimental results on the TriviaQA, MSMARCO, and CMRC2018 datasets demonstrate GenKI's effectiveness compared to state-of-the-art baselines, revealing a linear relationship between knowledge retrieval frequency and the model's ability to accurately retrieve knowledge.

Takeaways, Limitations

Takeaways:
A novel framework for simultaneously improving knowledge integration and controllable generation in LLM-based OpenQA is presented.
Achieving SOTA performance on various datasets (TriviaQA, MSMARCO, CMRC2018)
Establishing a linear relationship between knowledge retrieval frequency and accurate knowledge retrieval ability.
Limitations:
Specific Limitations information is not stated in the abstract (you will need to read the paper in detail later)
👍