Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

TaylorPODA: A Taylor Expansion-Based Method to Improve Post-Hoc Attributions for Opaque Models

Created by
  • Haebom

Author

Yuchi Tang, I naki Esnaola, George Panoutsos

Outline

Existing model-agnostic posterior explanation methods generate external explanations for opaque models primarily by locally attributing model outputs to input features. However, they lack a framework that explicitly and systematically quantifies the contributions of individual features. This paper integrates existing local attribution methods based on the Taylor expansion framework proposed by Deng et al. (2024) and presents strict assumptions for Taylor-specific attribution: precision, association, and zero-discrepancy. Building on these assumptions, we propose TaylorPODA (Taylor expansion-derived imPortance-Order aDapted Attribution), which incorporates an additional "adaptive" property. This property enables alignment with task-specific objectives, particularly in posterior settings where ground-truth explanations are lacking. Experimental evaluations demonstrate that TaylorPODA achieves competitive results compared to baseline methods and provides principled and easily visualized explanations. This study enhances the reliable distribution of opaque models by providing explanations with a stronger theoretical foundation.

Takeaways, Limitations

Takeaways:
Integrating existing local attribution methods and strengthening the theoretical foundation using the Taylor expansion framework.
Provides a systematic framework for Taylor-term specific attribution by proposing strict assumptions on "precision," "coalition," and "zero-discrepancy."
Additional “adaptive” properties allow for generating explanations tailored to task-specific goals.
It shows competitive performance compared to existing methods and provides a principled and easy-to-visualize explanation.
Contributes to improving the reliable distribution of opaque models.
Limitations:
Further research is needed to determine the generality and scope of applicability of the presented assumptions.
There is no guarantee that TaylorPODA will perform well in all situations, and performance may vary depending on specific datasets or models.
Since the quality of the explanation can vary depending on how the "adaptation" property is implemented, choosing an appropriate adaptation strategy is important.
Further validation of applicability and efficiency for high-dimensional data is needed.
👍