Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

DistShap: Scalable GNN Explanations with Distributed Shapley Values

Created by
  • Haebom

Author

Selahattin Akkas, Aditya Devarakonda, Ariful Azad

Outline

This paper addresses the problem of explaining the prediction results of graph neural networks (GNNs). It points out that attributing the predictions of GNNs to specific edges or features is computationally very expensive, and especially emphasizes the difficulty of identifying important edges among millions of candidate edges during the classification process of nodes with many neighbors. To address this problem, this paper proposes DistShap, a parallel algorithm that distributes Shapley value-based explanations across multiple GPUs. DistShap samples subgraphs in a distributed environment, runs GNN inference in parallel on GPUs, and computes edge importance scores by solving a distributed least-squares problem. Experimental results show that DistShap outperforms most existing GNN explanation methods and scales to GNN models with millions of features using up to 128 GPUs on the NERSC Perlmutter supercomputer.

Takeaways, Limitations

Takeaways:
The scalability problem of the Shapley value-based GNN explanation method was solved.
We enable explanation of large-scale GNN models by leveraging distributed computing.
Provides GNN prediction explanations with higher accuracy than existing methods.
By enabling explanations for GNN models with millions of features, we increase their applicability to real-world large-scale datasets.
Limitations:
DistShap's performance may depend on the number of GPUs available.
Due to the inherent complexity of computing Shapley values, description of very large graphs can still be computationally difficult.
It may only show optimal performance for certain types of GNN models or graph structures.
The accuracy of the explanation may be affected by the subgraph sampling strategy.
👍