Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Do Sparse Subnetworks Exhibit Cognitively Aligned Attention? Effects of Pruning on Saliency Map Fidelity, Sparsity, and Concept Coherence

Created by
  • Haebom

Author

Sanish Suwal, Dipkamal Bhusal, Michael Clifford, Nidhi Rastogi

Outline

We study the impact of pruning to maintain neural network performance on model interpretability. We investigate the impact of fine-tuning after size-based pruning on low-level importance maps and high-level concept representations. Using ResNet-18 trained on the ImageNette dataset, we compare importance maps at different pruning levels using Vanilla Gradients (VG) and Integrated Gradients (IG), assessing sparsity and fidelity. We track changes in semantic consistency of learned concepts using CRAFT-based concept extraction. Light pruning improves the focus and fidelity of the importance map and preserves semantically meaningful concepts. Heavy pruning reduces importance map sparsity and concept coherence while maintaining accuracy.

Takeaways, Limitations

Takeaways:
Light pruning can improve interpretability.
Excessive pruning hinders interpretability.
Pruning can shape internal representations into human-centered attentional patterns.
Limitations:
Limitations of the study model and dataset.
Limitations of interpretability metrics.
Further research is needed to determine generalizability.
👍