Rigorous Feature Importance Scores based on Shapley Value and Banzhaf Index
- URL: http://arxiv.org/abs/2508.11959v1
- Date: Sat, 16 Aug 2025 07:41:25 GMT
- Title: Rigorous Feature Importance Scores based on Shapley Value and Banzhaf Index
- Authors: Xuanxiang Huang, Olivier Létoffé, Joao Marques-Silva,
- Abstract summary: This paper proposes two novel feature importance scores based on Shapley value and Banzhaf index.<n>We take into account non-WAXp sets when computing feature contribution, and the novel scores quantify how effective each feature is at excluding AExs.
- Score: 3.3766484312332303
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Feature attribution methods based on game theory are ubiquitous in the field of eXplainable Artificial Intelligence (XAI). Recent works proposed rigorous feature attribution using logic-based explanations, specifically targeting high-stakes uses of machine learning (ML) models. Typically, such works exploit weak abductive explanation (WAXp) as the characteristic function to assign importance to features. However, one possible downside is that the contribution of non-WAXp sets is neglected. In fact, non-WAXp sets can also convey important information, because of the relationship between formal explanations (XPs) and adversarial examples (AExs). Accordingly, this paper leverages Shapley value and Banzhaf index to devise two novel feature importance scores. We take into account non-WAXp sets when computing feature contribution, and the novel scores quantify how effective each feature is at excluding AExs. Furthermore, the paper identifies properties and studies the computational complexity of the proposed scores.
Related papers
- Outcome-Based Online Reinforcement Learning: Algorithms and Fundamental Limits [58.63897489864948]
Reinforcement learning with outcome-based feedback faces a fundamental challenge.<n>How do we assign credit to the right actions?<n>This paper provides the first comprehensive analysis of this problem in online RL with general function approximation.
arXiv Detail & Related papers (2025-05-26T17:44:08Z) - From SHAP Scores to Feature Importance Scores [4.8158930873043335]
This paper shows that there is an essential relationship between feature attribution and a priori voting power.
It remains unclear how some of the most widely used power indices might be exploited as feature importance scores (FISs) in XAI.
arXiv Detail & Related papers (2024-05-20T03:52:41Z) - RankingSHAP -- Listwise Feature Attribution Explanations for Ranking Models [48.895510739010355]
We present three key contributions to address this gap.<n>First, we rigorously define listwise feature attribution for ranking models.<n>Second, we introduce RankingSHAP, extending the popular SHAP framework to accommodate listwise ranking attribution.<n>Third, we propose two novel evaluation paradigms for assessing the faithfulness of attributions in learning-to-rank models.
arXiv Detail & Related papers (2024-03-24T10:45:55Z) - Anytime Approximate Formal Feature Attribution [33.195028992904355]
Key explainability question is: given this decision was made, what are the input features which contributed to the decision?
Heuristic XAI approaches suffer from the lack of quality guarantees, and often try to approximate Shapley values, which is not the same as explaining which features contribute to a decision.
A recent alternative is so-called formal feature attribution (FFA), which defines feature importance as the fraction of formal abductive explanations (AXp's) containing the given feature.
arXiv Detail & Related papers (2023-12-12T04:24:05Z) - Fast Shapley Value Estimation: A Unified Approach [71.92014859992263]
We propose a straightforward and efficient Shapley estimator, SimSHAP, by eliminating redundant techniques.
In our analysis of existing approaches, we observe that estimators can be unified as a linear transformation of randomly summed values from feature subsets.
Our experiments validate the effectiveness of our SimSHAP, which significantly accelerates the computation of accurate Shapley values.
arXiv Detail & Related papers (2023-11-02T06:09:24Z) - On Computing Probabilistic Abductive Explanations [30.325691263226968]
The most widely studied explainable AI (XAI) approaches are unsound.
PI-explanations also exhibit important drawbacks, the most visible of which is arguably their size.
This paper investigates practical approaches for computing relevant sets for a number of widely used classifiers.
arXiv Detail & Related papers (2022-12-12T15:47:10Z) - On Computing Relevant Features for Explaining NBCs [5.71097144710995]
It is the case that modelagnostic explainable AI (XAI) can produce incorrect explanations.
PI-explanations also exhibit important drawbacks, the most visible of which is arguably their size.
This paper investigates the complexity of computing sets of relevant features for Naive Bayes classifiers (NBCs) and shows that, in practice, these are easy to compute.
arXiv Detail & Related papers (2022-07-11T10:12:46Z) - Predicting is not Understanding: Recognizing and Addressing
Underspecification in Machine Learning [47.651130958272155]
Underspecification refers to the existence of multiple models that are indistinguishable in their in-domain accuracy.
We formalize the concept of underspecification and propose a method to identify and partially address it.
arXiv Detail & Related papers (2022-07-06T11:20:40Z) - Explainability in Process Outcome Prediction: Guidelines to Obtain
Interpretable and Faithful Models [77.34726150561087]
We define explainability through the interpretability of the explanations and the faithfulness of the explainability model in the field of process outcome prediction.
This paper contributes a set of guidelines named X-MOP which allows selecting the appropriate model based on the event log specifications.
arXiv Detail & Related papers (2022-03-30T05:59:50Z) - Learning Aggregation Functions [78.47770735205134]
We introduce LAF (Learning Aggregation Functions), a learnable aggregator for sets of arbitrary cardinality.
We report experiments on semi-synthetic and real data showing that LAF outperforms state-of-the-art sum- (max-) decomposition architectures.
arXiv Detail & Related papers (2020-12-15T18:28:53Z) - On the Tractability of SHAP Explanations [40.829629145230356]
SHAP explanations are a popular feature-attribution mechanism for explainable AI.
We show that the complexity of computing the SHAP explanation is the same as the complexity of computing the expected value of the model.
Going beyond fully-factorized distributions, we show that computing SHAP explanations is already intractable for a very simple setting.
arXiv Detail & Related papers (2020-09-18T05:48:15Z) - Explaining Black Box Predictions and Unveiling Data Artifacts through
Influence Functions [55.660255727031725]
Influence functions explain the decisions of a model by identifying influential training examples.
We conduct a comparison between influence functions and common word-saliency methods on representative tasks.
We develop a new measure based on influence functions that can reveal artifacts in training data.
arXiv Detail & Related papers (2020-05-14T00:45:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.