Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Learning to Diagnose Privately: DP-Powered LLMs for Radiology Report Classification

Created by
  • Haebom

Author

Payel Bhattacharjee, Fengwei Tian, Geoffrey D. Rubin, Joseph Y. Lo, Nirav Merchant, Heidi Hanson, John Gounley, Ravi Tandon

Outline

This study proposes a framework for fine-tuning a large-scale language model (LLM) using differential privacy (DP) for multi-object classification using medical image report text. By injecting compensated noise during the fine-tuning process, we aim to mitigate privacy risks associated with sensitive patient data and prevent data leakage while maintaining classification performance. Using 50,232 medical image reports from publicly available MIMIC-CXR chest radiograph and CT-RATE computed tomography datasets collected from 2011 to 2019, we fine-tuned the LLM using differential privacy low-dimensional adaptation (DP-LoRA) on three model architectures: BERT-medium, BERT-small, and ALBERT-base, to classify 14 labels from the MIMIC-CXR dataset and 18 labels from the CT-RATE dataset. We evaluated model performance using the weighted F1 score at various privacy levels (privacy budget = {0.01, 0.1, 1.0, 10.0}) and compared model performance across different privacy levels to quantify the privacy-utility tradeoff. Experimental results revealed a clear privacy-utility tradeoff across two different datasets and three different models. Under moderate privacy guarantees, the DP fine-tuned model achieved weighted F1 scores of 0.88 for MIMIC-CXR and 0.59 for CT-RATE, demonstrating relatively similar performance to the non-privacy-preserving LoRA baseline models (0.90 and 0.78, respectively). In conclusion, differential privacy-preserving fine-tuning using LoRA enables effective and privacy-preserving multi-disease classification, addressing key challenges of LLM fine-tuning on sensitive medical data.

Takeaways, Limitations

Takeaways:
We present the possibility of simultaneously achieving privacy protection of medical data and improved LLM performance by utilizing differential privacy (DP) techniques.
An efficient DP-based LLM fine-tuning method using LoRA is presented.
Generalizability was verified through experiments on various medical datasets and model architectures.
Maintains relatively high classification performance even at intermediate privacy levels.
Limitations:
There is a trade-off between privacy level and model performance (privacy-utility trade-off).
Performance differences may occur depending on the characteristics of the dataset used (performance degradation of the CT-RATE dataset compared to MIMIC-CXR).
Further experiments are needed on more diverse medical datasets and LLM architectures.
Further validation is needed for application in actual clinical settings.
👍