Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Hallucination to Truth: A Review of Fact-Checking and Factuality Evaluation in Large Language Models

Created by
  • Haebom

Author

Subhey Sadi Rahman, Md. Adnanul Islam, Md. Mahbub Alam, Musarrat Zeba, Md. Abdur Rahman, Sadia Sultana Chowa, Mohaimenul Azam Khan Raiaan, Sami Azam

Outline

Large-Scale Language Models (LLMs) are trained on vast amounts of internet data containing inaccurate content, potentially generating misinformation. This review systematically analyzes methods for assessing the factual accuracy of LLM-generated content. It addresses key challenges, such as hallucinations, dataset limitations, and the reliability of evaluation metrics, and highlights the need for a robust fact-checking framework that integrates advanced prompting strategies, domain-specific fine-tuning, and augmented generation (RAG) methods. The current literature from 2020 to 2025 focuses on evaluation methods and mitigation techniques, addressing five research questions. Furthermore, it examines RAG frameworks for instruction tuning, multi-agent inference, and access to external knowledge.

Takeaways, Limitations

Current indicator Limitations
The importance of verified external evidence
Improving factual consistency through domain-specific customization
The importance of building more accurate, understandable, and context-aware fact-checking
Reliability issues of data sets and evaluation metrics
Further research is needed to assess the factual accuracy of LLM-generated content.
👍