Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Training with Explanations Alone: A New Paradigm to Prevent Shortcut Learning

Created by
  • Haebom

Author

Pedro RAS Bassi, Haydr AH Ali, Andrea Cavalli, Sergio Decherchi

Outline

This paper proposes Training with Explanations Alone (TEA), a novel learning paradigm, to address the problem of shortcut learning, which hinders the application of artificial intelligence (AI) in critical areas such as healthcare. TEA trains a classifier (a TEA student model) using the explanation heatmap of a teacher model as a target heatmap. This ensures that the TEA student model focuses on the same image features as the teacher model. Furthermore, if the teacher model is trained to ignore background bias, such as by removing the background, the student model is also trained to ignore background bias. By using multiple teacher models, the student model can be trained to be highly resistant to foreground bias, and surprisingly, it can achieve results that are consistent with the teacher model's output without applying a loss function to the student model's output. Comparing with 14 state-of-the-art methods on five datasets with strong background or foreground bias (including the Waterbirds and Xline datasets for COVID-19/pneumonia classification), the TEA student model demonstrates excellent resistance to bias, outperforms the state-of-the-art methods, and generalizes well to untrained hospital data.

Takeaways, Limitations

Takeaways:
A new approach to solving the shortcut learning problem in medical AI.
Suggesting the possibility of developing AI models with strong resistance to background and foreground biases.
Contributing to the development of explainable AI (XAI) technology
Increased data efficiency by enabling learning using only the teacher model's explanatory heatmap.
Improved generalization performance for diverse hospital data
Limitations:
The effectiveness of the proposed method may be limited to specific datasets.
There are aspects that depend on the performance of the teacher model.
Generalization performance verification is needed for complex medical image data.
Further research is needed to determine its applicability in real-world clinical settings.
👍