Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

MedQARo: A Large-Scale Benchmark for Medical Question Answering in Romanian

Created by
  • Haebom

Author

Ana-Cristina Rogoz, Radu Tudor Ionescu, Alexandra-Valentina Anghel, Ionut-Lucian Antone-Iordache, Simona Coniac, Andreea Iuliana Ionescu

Outline

This paper introduces MedQARo, a large-scale benchmark for medical question answering (QA) in the Romanian language. It presents a high-quality dataset consisting of 102,646 question-answer pairs related to medical case summaries of cancer patients, manually annotated by seven oncologists or radiologists over 2,100 hours of work. We evaluate four large-scale language models (LLMs) from different model families on MedQARo, comparing their performance under two scenarios: zero-shot prompting and supervised learning-based fine-tuning. We demonstrate that the fine-tuned models significantly outperform the zero-shot models, while pretrained models struggle to generalize on MedQARo. This study highlights the importance of domain-specific and language-specific fine-tuning in Romanian medical QA, and we make the MedQARo dataset and code publicly available.

Takeaways, Limitations

Takeaways:
We provide MedQARo, a large-scale, high-quality dataset in the field of Romanian medical QA.
We demonstrate the importance of fine-tuning for Romanian medical QA by comparing the performance of different LLMs.
Demonstrates the need for domain-specific and language-specific fine-tuning.
You can contribute to future research through open datasets and code.
Limitations:
The dataset is limited to cancer patients, so further research is needed to determine its generalizability to general healthcare settings.
The types of LLMs used in the evaluation may be limited.
There are cost and time constraints associated with manual annotation.
👍