groupShapley: Efficient prediction explanation with Shapley values for
feature groups
- URL: http://arxiv.org/abs/2106.12228v1
- Date: Wed, 23 Jun 2021 08:16:14 GMT
- Title: groupShapley: Efficient prediction explanation with Shapley values for
feature groups
- Authors: Martin Jullum, Annabelle Redelmeier, Kjersti Aas
- Abstract summary: Shapley values has established itself as one of the most appropriate and theoretically sound frameworks for explaining predictions from machine learning models.
The main drawback with Shapley values is that its computational complexity grows exponentially in the number of input features.
The present paper introduces groupShapley: a conceptually simple approach for dealing with the aforementioned bottlenecks.
- Score: 2.320417845168326
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Shapley values has established itself as one of the most appropriate and
theoretically sound frameworks for explaining predictions from complex machine
learning models. The popularity of Shapley values in the explanation setting is
probably due to its unique theoretical properties. The main drawback with
Shapley values, however, is that its computational complexity grows
exponentially in the number of input features, making it unfeasible in many
real world situations where there could be hundreds or thousands of features.
Furthermore, with many (dependent) features, presenting/visualizing and
interpreting the computed Shapley values also becomes challenging. The present
paper introduces groupShapley: a conceptually simple approach for dealing with
the aforementioned bottlenecks. The idea is to group the features, for example
by type or dependence, and then compute and present Shapley values for these
groups instead of for all individual features. Reducing hundreds or thousands
of features to half a dozen or so, makes precise computations practically
feasible and the presentation and knowledge extraction greatly simplified. We
prove that under certain conditions, groupShapley is equivalent to summing the
feature-wise Shapley values within each feature group. Moreover, we provide a
simulation study exemplifying the differences when these conditions are not
met. We illustrate the usability of the approach in a real world car insurance
example, where groupShapley is used to provide simple and intuitive
explanations.
Related papers
- Improving the Sampling Strategy in KernelSHAP [0.8057006406834466]
KernelSHAP framework enables us to approximate the Shapley values using a sampled subset of weighted conditional expectations.
We propose three main novel contributions: a stabilizing technique to reduce the variance of the weights in the current state-of-the-art strategy, a novel weighing scheme that corrects the Shapley kernel weights based on sampled subsets, and a straightforward strategy that includes the important subsets and integrates them with the corrected Shapley kernel weights.
arXiv Detail & Related papers (2024-10-07T10:02:31Z) - Shapley Pruning for Neural Network Compression [63.60286036508473]
This work presents the Shapley value approximations, and performs the comparative analysis in terms of cost-benefit utility for the neural network compression.
The proposed normative ranking and its approximations show practical results, obtaining state-of-the-art network compression.
arXiv Detail & Related papers (2024-07-19T11:42:54Z) - Fast Shapley Value Estimation: A Unified Approach [71.92014859992263]
We propose a straightforward and efficient Shapley estimator, SimSHAP, by eliminating redundant techniques.
In our analysis of existing approaches, we observe that estimators can be unified as a linear transformation of randomly summed values from feature subsets.
Our experiments validate the effectiveness of our SimSHAP, which significantly accelerates the computation of accurate Shapley values.
arXiv Detail & Related papers (2023-11-02T06:09:24Z) - Shapley Sets: Feature Attribution via Recursive Function Decomposition [6.85316573653194]
We propose an alternative attribution approach, Shapley Sets, which awards value to sets of features.
Shapley Sets decomposes the underlying model into non-separable variable groups.
We show theoretically and experimentally how Shapley Sets avoids pitfalls associated with Shapley value based alternatives.
arXiv Detail & Related papers (2023-07-04T15:30:09Z) - Efficient Shapley Values Estimation by Amortization for Text
Classification [66.7725354593271]
We develop an amortized model that directly predicts each input feature's Shapley Value without additional model evaluations.
Experimental results on two text classification datasets demonstrate that our amortized model estimates Shapley Values accurately with up to 60 times speedup.
arXiv Detail & Related papers (2023-05-31T16:19:13Z) - WeightedSHAP: analyzing and improving Shapley based feature attributions [17.340091573913316]
Shapley value is a popular approach for measuring the influence of individual features.
We propose WeightedSHAP, which generalizes the Shapley value and learns which marginal contributions to focus directly from data.
On several real-world datasets, we demonstrate that the influential features identified by WeightedSHAP are better able to recapitulate the model's predictions.
arXiv Detail & Related papers (2022-09-27T14:34:07Z) - An Additive Instance-Wise Approach to Multi-class Model Interpretation [53.87578024052922]
Interpretable machine learning offers insights into what factors drive a certain prediction of a black-box system.
Existing methods mainly focus on selecting explanatory input features, which follow either locally additive or instance-wise approaches.
This work exploits the strengths of both methods and proposes a global framework for learning local explanations simultaneously for multiple target classes.
arXiv Detail & Related papers (2022-07-07T06:50:27Z) - Exact Shapley Values for Local and Model-True Explanations of Decision
Tree Ensembles [0.0]
We consider the application of Shapley values for explaining decision tree ensembles.
We present a novel approach to Shapley value-based feature attribution that can be applied to random forests and boosted decision trees.
arXiv Detail & Related papers (2021-12-16T20:16:02Z) - Fast Hierarchical Games for Image Explanations [78.16853337149871]
We present a model-agnostic explanation method for image classification based on a hierarchical extension of Shapley coefficients.
Unlike other Shapley-based explanation methods, h-Shap is scalable and can be computed without the need of approximation.
We compare our hierarchical approach with popular Shapley-based and non-Shapley-based methods on a synthetic dataset, a medical imaging scenario, and a general computer vision problem.
arXiv Detail & Related papers (2021-04-13T13:11:02Z) - Causal Shapley Values: Exploiting Causal Knowledge to Explain Individual
Predictions of Complex Models [6.423239719448169]
Shapley values are designed to attribute the difference between a model's prediction and an average baseline to the different features used as input to the model.
We show how these 'causal' Shapley values can be derived for general causal graphs without sacrificing any of their desirable properties.
arXiv Detail & Related papers (2020-11-03T11:11:36Z) - Towards Efficient Data Valuation Based on the Shapley Value [65.4167993220998]
We study the problem of data valuation by utilizing the Shapley value.
The Shapley value defines a unique payoff scheme that satisfies many desiderata for the notion of data value.
We propose a repertoire of efficient algorithms for approximating the Shapley value.
arXiv Detail & Related papers (2019-02-27T00:22:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.