Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Knowledge-Level Consistency Reinforcement Learning: Dual-Fact Alignment for Long-Form Factuality

Created by
  • Haebom

Author

Junliang Li, Yucheng Wang, Yan Chen, Yu Ran, Ruiqing Zhang, Jing Liu, Hua Wu, Haifeng Wang

Outline

To address the problems of hallucination and lack of factual accuracy in long-form text generation from large-scale language models (LLMs), we propose the Knowledge-Level Consistency Reinforcement Learning Framework (KLCF). KLCF introduces a Dual-Fact Alignment mechanism that improves factual recall and precision by considering the model's internal knowledge boundaries. It utilizes pre-trained knowledge boundaries to construct a fact checklist and trains a self-evaluation module based on the base model's internal knowledge to enhance factual accuracy. Its lightweight and efficient reward design, requiring no external knowledge, makes it easily applicable to large-scale learning. Experimental results demonstrate that KLCF significantly improves factual accuracy in several long-form benchmarks and effectively mitigates model hallucination.

Takeaways, Limitations

Takeaways:
A new framework is presented to address the hallucination problem in LLM.
Development of a Dual-Fact Alignment mechanism that enhances factuality by leveraging internal knowledge boundaries.
Efficient and scalable reward design without external knowledge.
Demonstrated improvement in realism indicators and hallucination mitigation effects in several benchmarks.
Limitations:
Additional information is needed on how to build specific knowledge boundaries and train self-assessment modules.
Generalizability to other LLM architectures and tasks needs to be verified.
Computational complexity and training time analysis required.
👍