Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

MedShift: Implicit Conditional Transport for X-Ray Domain Adaptation

Created by
  • Haebom

Author

Francisco Caetano, Christiaan Viviers, Peter HH de With, Fons van der Sommen

Outline

Synthetic medical data offers a scalable solution for training robust models, but significant domain gaps limit generalization to real-world clinical settings. This paper addresses the problem of cross-domain translation between synthetic and real-world head X images by focusing on resolving discrepancies in attenuation behavior, noise characteristics, and soft tissue representation. We propose MedShift, a unified class-conditional generative model based on Flow Matching and Schrodinger Bridges. This model enables high-fidelity, unpaired image translation across multiple domains. Unlike previous approaches that require domain-specific training or rely on paired data, MedShift learns a shared, domain-independent latent space and enables seamless translation between all observed domain pairs during training. We also benchmark domain translation models by introducing X-DigiSkull, a novel dataset consisting of aligned synthetic and real skull X lines from various radiation doses. Experimental results demonstrate that MedShift delivers robust performance despite a smaller model size compared to diffusion-based approaches, and that it can be tuned to prioritize either perceptual fidelity or structural consistency during inference, making it a scalable and generalizable solution for domain adaptation in medical imaging. The code and dataset are available at https://caetas.github.io/medshift.html .

Takeaways, Limitations

Takeaways:
High-fidelity, unpaired image transformation between synthetic and real medical images is possible through our proposed MedShift, a unified class conditional generative model based on Flow Matching and Schrodinger Bridges.
Shared domain-independent latent space learning without domain-specific training or dependence on paired data.
Flexibility is achieved by allowing the prioritization of either perceptual fidelity or structural consistency during inference.
Provides robust performance with smaller model sizes than diffusion-based approaches.
New dataset X-DigiSkull enables benchmarking of domain transformation models.
Providing scalable and generalizable domain-adapted solutions
Limitations:
Limitations is not explicitly mentioned in the paper. Further research is needed to further verify its generalization performance in real-world clinical settings and its applicability to various medical imaging modalities.
👍