Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Hallucinations and Key Information Extraction in Medical Texts: A Comprehensive Assessment of Open-Source Large Language Models

Created by
  • Haebom

Author

Anindya Bijoy Das, Shibbir Ahmed, Shahnewaz Karim Sakib

Outline

This paper investigates the effectiveness of open-source large-scale language models (LLMs) for extracting key events (reasons for admission, major in-hospital events, and important follow-up measures) from medical summaries, specifically discharge reports. We also evaluate the incidence of hallucinations, which can affect the accuracy and reliability of LLMs. Experiments using LLMs such as Qwen2.5 and DeepSeek-v2 demonstrate excellent performance in extracting reasons for admission and events occurring during hospitalization, but show inconsistencies in identifying follow-up recommendations. This highlights the challenges of leveraging LLMs for comprehensive summarization.

Takeaways, Limitations

Takeaways: We demonstrate that open-source LLM is effective in extracting medical summaries, particularly reasons for admission and critical incidents within the hospital. This suggests the potential for developing an LLM-based automated medical summarization system.
Limitations: LLMs have shown inconsistent extraction of certain information, such as follow-up recommendations. The potential for hallucinations in LLMs and the resulting reliability issues with medical information should be considered. Further research is needed to utilize LLMs for comprehensive medical summaries.
👍