SHAP-XRT: The Shapley Value Meets Conditional Independence Testing
- URL: http://arxiv.org/abs/2207.07038v5
- Date: Wed, 27 Dec 2023 15:58:39 GMT
- Title: SHAP-XRT: The Shapley Value Meets Conditional Independence Testing
- Authors: Jacopo Teneggi, Beepul Bharti, Yaniv Romano and Jeremias Sulam
- Abstract summary: We show that Shapley-based explanation methods and conditional independence testing are closely related.
We introduce the SHAPley EXplanation Randomization Test (SHAP-XRT), a testing procedure inspired by the Conditional Randomization Test (CRT) for a specific notion of local (i.e., on a sample) conditional independence.
We show that the Shapley value itself provides an upper bound to the expected $p$-value of a global (i.e., overall) null hypothesis.
- Score: 21.794110108580746
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The complex nature of artificial neural networks raises concerns on their
reliability, trustworthiness, and fairness in real-world scenarios. The Shapley
value -- a solution concept from game theory -- is one of the most popular
explanation methods for machine learning models. More traditionally, from a
statistical perspective, feature importance is defined in terms of conditional
independence. So far, these two approaches to interpretability and feature
importance have been considered separate and distinct. In this work, we show
that Shapley-based explanation methods and conditional independence testing are
closely related. We introduce the SHAPley EXplanation Randomization Test
(SHAP-XRT), a testing procedure inspired by the Conditional Randomization Test
(CRT) for a specific notion of local (i.e., on a sample) conditional
independence. With it, we prove that for binary classification problems, the
marginal contributions in the Shapley value provide lower and upper bounds to
the expected $p$-values of their respective tests. Furthermore, we show that
the Shapley value itself provides an upper bound to the expected $p$-value of a
global (i.e., overall) null hypothesis. As a result, we further our
understanding of Shapley-based explanation methods from a novel perspective and
characterize the conditions under which one can make statistically valid claims
about feature importance via the Shapley value.
Related papers
- Precision of Individual Shapley Value Explanations [0.0]
Shapley values are extensively used in explainable artificial intelligence (XAI) as a framework to explain predictions made by complex machine learning (ML) models.
We show that the explanations are systematically less precise for observations on the outer region of the training data distribution.
This is expected from a statistical point of view, but to the best of our knowledge, it has not been systematically addressed in the Shapley value literature.
arXiv Detail & Related papers (2023-12-06T13:29:23Z) - Fast Shapley Value Estimation: A Unified Approach [71.92014859992263]
We propose a straightforward and efficient Shapley estimator, SimSHAP, by eliminating redundant techniques.
In our analysis of existing approaches, we observe that estimators can be unified as a linear transformation of randomly summed values from feature subsets.
Our experiments validate the effectiveness of our SimSHAP, which significantly accelerates the computation of accurate Shapley values.
arXiv Detail & Related papers (2023-11-02T06:09:24Z) - Efficient Shapley Values Estimation by Amortization for Text
Classification [66.7725354593271]
We develop an amortized model that directly predicts each input feature's Shapley Value without additional model evaluations.
Experimental results on two text classification datasets demonstrate that our amortized model estimates Shapley Values accurately with up to 60 times speedup.
arXiv Detail & Related papers (2023-05-31T16:19:13Z) - Shortcomings of Top-Down Randomization-Based Sanity Checks for
Evaluations of Deep Neural Network Explanations [67.40641255908443]
We identify limitations of model-randomization-based sanity checks for the purpose of evaluating explanations.
Top-down model randomization preserves scales of forward pass activations with high probability.
arXiv Detail & Related papers (2022-11-22T18:52:38Z) - Nonparametric Conditional Local Independence Testing [69.31200003384122]
Conditional local independence is an independence relation among continuous time processes.
No nonparametric test of conditional local independence has been available.
We propose such a nonparametric test based on double machine learning.
arXiv Detail & Related papers (2022-03-25T10:31:02Z) - Is Shapley Explanation for a model unique? [0.0]
We explore the relationship between the distribution of a feature and its Shapley value.
Our assessment is that Shapley value for particular feature not only depends on its expected mean but on other moments as well such as variance.
It varies with model outcome (Probability/Log-odds/binary decision such as accept vs reject) and hence model application.
arXiv Detail & Related papers (2021-11-23T15:31:46Z) - Collective eXplainable AI: Explaining Cooperative Strategies and Agent
Contribution in Multiagent Reinforcement Learning with Shapley Values [68.8204255655161]
This study proposes a novel approach to explain cooperative strategies in multiagent RL using Shapley values.
Results could have implications for non-discriminatory decision making, ethical and responsible AI-derived decisions or policy making under fairness constraints.
arXiv Detail & Related papers (2021-10-04T10:28:57Z) - groupShapley: Efficient prediction explanation with Shapley values for
feature groups [2.320417845168326]
Shapley values has established itself as one of the most appropriate and theoretically sound frameworks for explaining predictions from machine learning models.
The main drawback with Shapley values is that its computational complexity grows exponentially in the number of input features.
The present paper introduces groupShapley: a conceptually simple approach for dealing with the aforementioned bottlenecks.
arXiv Detail & Related papers (2021-06-23T08:16:14Z) - Causal Shapley Values: Exploiting Causal Knowledge to Explain Individual
Predictions of Complex Models [6.423239719448169]
Shapley values are designed to attribute the difference between a model's prediction and an average baseline to the different features used as input to the model.
We show how these 'causal' Shapley values can be derived for general causal graphs without sacrificing any of their desirable properties.
arXiv Detail & Related papers (2020-11-03T11:11:36Z) - Predictive and Causal Implications of using Shapley Value for Model
Interpretation [6.744385328015561]
We established the relationship between Shapley value and conditional independence, a key concept in both predictive and causal modeling.
Our results indicate that, eliminating a variable with high Shapley value from a model do not necessarily impair predictive performance.
More importantly, Shapley value of a variable do not reflect their causal relationship with the target of interest.
arXiv Detail & Related papers (2020-08-12T01:08:08Z) - Towards Efficient Data Valuation Based on the Shapley Value [65.4167993220998]
We study the problem of data valuation by utilizing the Shapley value.
The Shapley value defines a unique payoff scheme that satisfies many desiderata for the notion of data value.
We propose a repertoire of efficient algorithms for approximating the Shapley value.
arXiv Detail & Related papers (2019-02-27T00:22:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.