A $k$-additive Choquet integral-based approach to approximate the SHAP
values for local interpretability in machine learning
- URL: http://arxiv.org/abs/2211.02166v1
- Date: Thu, 3 Nov 2022 22:34:50 GMT
- Title: A $k$-additive Choquet integral-based approach to approximate the SHAP
values for local interpretability in machine learning
- Authors: Guilherme Dean Pelegrina, Leonardo Tomazeli Duarte, Michel Grabisch
- Abstract summary: This paper aims at providing some interpretability for machine learning models based on Shapley values.
A SHAP-based method called Kernel SHAP adopts an efficient strategy that approximates such values with less computational effort.
The obtained results attest that our proposal needs less computations on coalitions of attributes to approximate the SHAP values.
- Score: 8.637110868126546
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Besides accuracy, recent studies on machine learning models have been
addressing the question on how the obtained results can be interpreted. Indeed,
while complex machine learning models are able to provide very good results in
terms of accuracy even in challenging applications, it is difficult to
interpret them. Aiming at providing some interpretability for such models, one
of the most famous methods, called SHAP, borrows the Shapley value concept from
game theory in order to locally explain the predicted outcome of an instance of
interest. As the SHAP values calculation needs previous computations on all
possible coalitions of attributes, its computational cost can be very high.
Therefore, a SHAP-based method called Kernel SHAP adopts an efficient strategy
that approximate such values with less computational effort. In this paper, we
also address local interpretability in machine learning based on Shapley
values. Firstly, we provide a straightforward formulation of a SHAP-based
method for local interpretability by using the Choquet integral, which leads to
both Shapley values and Shapley interaction indices. Moreover, we also adopt
the concept of $k$-additive games from game theory, which contributes to reduce
the computational effort when estimating the SHAP values. The obtained results
attest that our proposal needs less computations on coalitions of attributes to
approximate the SHAP values.
Related papers
- Improving the Sampling Strategy in KernelSHAP [0.8057006406834466]
KernelSHAP framework enables us to approximate the Shapley values using a sampled subset of weighted conditional expectations.
We propose three main novel contributions: a stabilizing technique to reduce the variance of the weights in the current state-of-the-art strategy, a novel weighing scheme that corrects the Shapley kernel weights based on sampled subsets, and a straightforward strategy that includes the important subsets and integrates them with the corrected Shapley kernel weights.
arXiv Detail & Related papers (2024-10-07T10:02:31Z) - Provably Accurate Shapley Value Estimation via Leverage Score Sampling [12.201705893125775]
We introduce Leverage SHAP, a light-weight modification of Kernel SHAP that provides provably accurate Shapley value estimates with just $O(nlog n)$ model evaluations.
Our approach takes advantage of a connection between Shapley value estimation and active learning by employing leverage score sampling, a powerful regression tool.
arXiv Detail & Related papers (2024-10-02T18:15:48Z) - Energy-based Model for Accurate Shapley Value Estimation in Interpretable Deep Learning Predictive Modeling [7.378438977893025]
EmSHAP is an energy-based model for Shapley value estimation.
It estimates the expectation of Shapley contribution function under arbitrary subset of features.
arXiv Detail & Related papers (2024-04-01T12:19:33Z) - Fast Shapley Value Estimation: A Unified Approach [71.92014859992263]
We propose a straightforward and efficient Shapley estimator, SimSHAP, by eliminating redundant techniques.
In our analysis of existing approaches, we observe that estimators can be unified as a linear transformation of randomly summed values from feature subsets.
Our experiments validate the effectiveness of our SimSHAP, which significantly accelerates the computation of accurate Shapley values.
arXiv Detail & Related papers (2023-11-02T06:09:24Z) - Efficient Model-Free Exploration in Low-Rank MDPs [76.87340323826945]
Low-Rank Markov Decision Processes offer a simple, yet expressive framework for RL with function approximation.
Existing algorithms are either (1) computationally intractable, or (2) reliant upon restrictive statistical assumptions.
We propose the first provably sample-efficient algorithm for exploration in Low-Rank MDPs.
arXiv Detail & Related papers (2023-07-08T15:41:48Z) - Generalizing Backpropagation for Gradient-Based Interpretability [103.2998254573497]
We show that the gradient of a model is a special case of a more general formulation using semirings.
This observation allows us to generalize the backpropagation algorithm to efficiently compute other interpretable statistics.
arXiv Detail & Related papers (2023-07-06T15:19:53Z) - Shapley Computations Using Surrogate Model-Based Trees [4.2575268077562685]
This paper proposes the use of a surrogate model-based tree to compute Shapley and SHAP values based on conditional expectation.
Simulation studies show that the proposed algorithm provides improvements in accuracy, unifies global Shapley and SHAP interpretation, and the thresholding method provides a way to trade-off running time and accuracy.
arXiv Detail & Related papers (2022-07-11T22:20:51Z) - Accelerating Shapley Explanation via Contributive Cooperator Selection [42.11059072201565]
We propose a novel method SHEAR to significantly accelerate the Shapley explanation for DNN models.
The selection of the feature coalitions follows our proposed Shapley chain rule to minimize the absolute error from the ground-truth Shapley values.
SHEAR consistently outperforms state-of-the-art baseline methods across different evaluation metrics.
arXiv Detail & Related papers (2022-06-17T03:24:45Z) - Fast Hierarchical Games for Image Explanations [78.16853337149871]
We present a model-agnostic explanation method for image classification based on a hierarchical extension of Shapley coefficients.
Unlike other Shapley-based explanation methods, h-Shap is scalable and can be computed without the need of approximation.
We compare our hierarchical approach with popular Shapley-based and non-Shapley-based methods on a synthetic dataset, a medical imaging scenario, and a general computer vision problem.
arXiv Detail & Related papers (2021-04-13T13:11:02Z) - Efficient semidefinite-programming-based inference for binary and
multi-class MRFs [83.09715052229782]
We propose an efficient method for computing the partition function or MAP estimate in a pairwise MRF.
We extend semidefinite relaxations from the typical binary MRF to the full multi-class setting, and develop a compact semidefinite relaxation that can again be solved efficiently using the solver.
arXiv Detail & Related papers (2020-12-04T15:36:29Z) - Provably Efficient Reward-Agnostic Navigation with Linear Value
Iteration [143.43658264904863]
We show how iteration under a more standard notion of low inherent Bellman error, typically employed in least-square value-style algorithms, can provide strong PAC guarantees on learning a near optimal value function.
We present a computationally tractable algorithm for the reward-free setting and show how it can be used to learn a near optimal policy for any (linear) reward function.
arXiv Detail & Related papers (2020-08-18T04:34:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.