Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Who Does What in Deep Learning? Multidimensional Game-Theoretic Attribution of Function of Neural Units

Created by
  • Haebom

Author

Shrey Dixit, Kayson Fakhar, Fatemeh Hadaeghi, Patrick Mineault, Konrad P. Kording, Claus C. Hilgetag

Outline

In this paper, we present an explainable AI (XAI) method for how each neuron contributes to the output of neural networks with billions of parameters, such as large language models (LLMs) and generative adversarial networks (GANs). Existing XAI methods assign importance to inputs, but are unable to quantify the contribution of neurons across thousands of output pixels, tokens, or logits. We address this problem by presenting multi-perturbation Shapley value analysis (MSA), a model-agnostic game-theoretic framework. MSA systematically removes combinations of neurons to produce Shapley modes, which are per-unit contribution maps with the same dimension as the model output. We apply MSA to models of various sizes, from multilayer perceptrons to Mixtral-8x7B with 56 billion parameters and GANs, demonstrating how regularization concentrates computation on a small number of hubs, language-specific experts in LLMs, and inverted pixel-generating hierarchies in GANs. These results demonstrate that MSA is a powerful approach for interpreting, compiling, and compressing deep neural networks.

Takeaways, Limitations

Takeaways:
A novel method (MSA) to quantitatively analyze the contribution of individual neurons to the output of large-scale neural networks is presented.
MSA uncovers new insights from language-specific experts within LLM, including the inverted pixel generation hierarchy of GANs
Provides powerful tools for model interpretation, editing, and compression
Applicability to models of various sizes (MLP, Mixtral-8x7B, GAN) confirmed
Limitations:
MSA can be computationally complex (especially for very large models)
Further research is needed to determine whether MSA is applicable to all types of neural networks.
Additional guidance may be needed on interpreting MSA results.
👍