[공지사항]을 빙자한 안부와 근황 
Show more

Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

FluoroSAM: A Language-promptable Foundation Model for Flexible X-ray Image Segmentation

Created by
  • Haebom

Author

Benjamin D. Killeen, Liam J. Wang, Blanca Inigo, Han Zhang, Mehran Armand, Russell H. Taylor, Greg Osgood, Mathias Unberath

Outline

In this paper, we present FluoroSAM, a language-promptable Xline image segmentation model that can segment diverse human anatomical structures and instruments based on natural language prompts. While existing medical image analysis-based models require large-scale, richly annotated datasets, FluoroSAM is trained on an extensive dataset of 3 million synthetic Xline images containing diverse anatomical structures (128 organ types) and instruments (464 instruments). We improve the segmentation performance for natural language prompts by introducing vector quantization (VQ) of text embeddings, and demonstrate its rich support for human-machine interaction through quantitative performance evaluations on real Xline images and several application examples. The code is publicly available on GitHub.

Takeaways, Limitations

Takeaways:
We present a model-based framework for comprehensive and language-aligned analysis of a variety of X line images (from diagnostic chest X radiographs to interventional fluoroscopy).
Accurately segment various anatomical structures and tools using natural language prompts.
Medical __T70582_____Enables rich human-machine interaction in imaging acquisition and analysis environments.
Reduced the burden of annotating real data by leveraging 3 million synthetic __T70583_____ line image data.
Limitations:
Since it was trained using synthetic data, its generalization performance to real X line images may be limited.
Although it encompasses a wide range of Xline imaging types, it may perform worse for certain types of Xline imaging or diseases than others.
Additional real-world data and annotations may be required to further improve model performance.
👍