Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Exploring Scaling Laws for EHR Foundation Models

Created by
  • Haebom

Author

Sheng Zhang, Qin Liu, Naoto Usuyama, Cliff Wong, Tristan Naumann, Hoifung Poon

Outline

This paper presents the first empirical study to determine whether the scaling laws of large-scale language models (LLMs) can be applied to electronic health record (EHR)-based models. Using patient time-series data from the MIMIC-IV database, we trained Transformer architectures with various model sizes and compute budgets. We observed consistent scaling patterns, including a quadratic IsoFLOPs curve and a power-law relationship between computation, model parameters, data size, and clinical utility. This demonstrates that EHR models exhibit similar scaling behavior to LLMs, providing predictive insights for resource-efficient training strategies. Consequently, this study lays the foundation for developing robust EHR-based models that can transform clinical prediction tasks and advance personalized medicine.

Takeaways, Limitations

Takeaways:
We empirically confirmed that a similar scaling law to LLM exists in EHR-based models.
By elucidating the relationships between computing resources, model size, data size, and clinical utility, we contribute to the establishment of resource-efficient model training strategies.
It lays the foundation for developing robust EHR-based models that can contribute to the advancement of clinical prediction and personalized medicine.
Limitations:
Because the study was conducted using only one MIMIC-IV database, further research is needed to determine generalizability to other EHR datasets.
Since this study is limited to a specific architecture (Transformer), the applicability of the scaling law to other architectures needs to be verified.
It is necessary to secure diversity in clinical usefulness evaluation indicators and establish objective evaluation criteria.
👍