Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Leveraging Online Data to Enhance Medical Knowledge in a Small Persian Language Model

Created by
  • Haebom

Author

Mehrdad Ghassabi, Pedram Rostami, Hamidreza Baradaran Kashani, Amirhossein Poursina, Zahra Kazemi, Milad Tavakoli

Outline

To overcome the limitations of small language models in specialized fields such as Persian, a low-resource language, this study introduces a new dataset consisting of 20,000 doctor-patient question-and-answer pairs and a 90 million-token corpus crawled from medical journals. Using this dataset, we improved the medical knowledge of the baseline model, aya-expanse-8b, through parameter-efficient fine-tuning.

Takeaways, Limitations

Demonstrating the applicability of small language models to Persian medical fields.
Presenting medical AI solutions in resource-constrained environments using open-access online data.
Fine-tuned model improves medical question answering accuracy and passes IBSEE.
MMLU accuracy improved by an average of 2.67% when translated into Persian.
Future research could explore performance enhancements through multimodal input.
👍