Unifying Feature-Based Explanations with Functional ANOVA and Cooperative Game Theory
- URL: http://arxiv.org/abs/2412.17152v1
- Date: Sun, 22 Dec 2024 20:12:16 GMT
- Title: Unifying Feature-Based Explanations with Functional ANOVA and Cooperative Game Theory
- Authors: Fabian Fumagalli, Maximilian Muschalik, Eyke Hüllermeier, Barbara Hammer, Julia Herbinger,
- Abstract summary: We introduce a unified framework for local and global feature-based explanations.
We introduce three fANOVA decompositions that determine the influence of feature distributions.
We then empirically showcase the usefulness of our framework on synthetic and real-world datasets.
- Score: 21.405333804566357
- License:
- Abstract: Feature-based explanations, using perturbations or gradients, are a prevalent tool to understand decisions of black box machine learning models. Yet, differences between these methods still remain mostly unknown, which limits their applicability for practitioners. In this work, we introduce a unified framework for local and global feature-based explanations using two well-established concepts: functional ANOVA (fANOVA) from statistics, and the notion of value and interaction from cooperative game theory. We introduce three fANOVA decompositions that determine the influence of feature distributions, and use game-theoretic measures, such as the Shapley value and interactions, to specify the influence of higher-order interactions. Our framework combines these two dimensions to uncover similarities and differences between a wide range of explanation techniques for features and groups of features. We then empirically showcase the usefulness of our framework on synthetic and real-world datasets.
Related papers
- A Unifying Framework for Action-Conditional Self-Predictive Reinforcement Learning [48.59516337905877]
Learning a good representation is a crucial challenge for Reinforcement Learning (RL) agents.
Recent work has developed theoretical insights into these algorithms.
We take a step towards bridging the gap between theory and practice by analyzing an action-conditional self-predictive objective.
arXiv Detail & Related papers (2024-06-04T07:22:12Z) - On the Joint Interaction of Models, Data, and Features [82.60073661644435]
We introduce a new tool, the interaction tensor, for empirically analyzing the interaction between data and model through features.
Based on these observations, we propose a conceptual framework for feature learning.
Under this framework, the expected accuracy for a single hypothesis and agreement for a pair of hypotheses can both be derived in closed-form.
arXiv Detail & Related papers (2023-06-07T21:35:26Z) - The Weighted M\"obius Score: A Unified Framework for Feature Attribution [17.358276581599643]
Feature attribution aims to explain the reasoning behind a black-box model's prediction by identifying the impact of each feature on the prediction.
The lack of a unified framework has led to a proliferation of methods that are often not directly comparable.
This paper introduces a parameterized attribution framework -- the Weighted M"obius Score -- and shows that many different attribution methods for both individual features and feature interactions are special cases.
arXiv Detail & Related papers (2023-05-16T06:27:27Z) - Distributing Synergy Functions: Unifying Game-Theoretic Interaction
Methods for Machine-Learning Explainability [9.416757363901295]
We present a unifying framework for game-theory-inspired attribution and $ktextth$-order interaction methods.
We identify how various methods are characterized by their policy of distributing synergies.
We show that the combination of various criteria uniquely defines the attribution/interaction methods.
arXiv Detail & Related papers (2023-05-04T18:33:18Z) - Function Composition in Trustworthy Machine Learning: Implementation
Choices, Insights, and Questions [28.643482049799477]
This paper focuses on compositions of functions arising from the different 'pillars' of trustworthiness.
We report initial empirical results and new insights on 7 real-world trustworthy dimensions - fairness and explainability.
We also report progress, and implementation choices, on an composer tool to encourage the combination of functionalities from multiple pillars.
arXiv Detail & Related papers (2023-02-17T23:49:16Z) - Relational Local Explanations [11.679389861042]
We develop a novel model-agnostic and permutation-based feature attribution algorithm based on relational analysis between input variables.
We are able to gain a broader insight into machine learning model decisions and data.
arXiv Detail & Related papers (2022-12-23T14:46:23Z) - MACE: An Efficient Model-Agnostic Framework for Counterfactual
Explanation [132.77005365032468]
We propose a novel framework of Model-Agnostic Counterfactual Explanation (MACE)
In our MACE approach, we propose a novel RL-based method for finding good counterfactual examples and a gradient-less descent method for improving proximity.
Experiments on public datasets validate the effectiveness with better validity, sparsity and proximity.
arXiv Detail & Related papers (2022-05-31T04:57:06Z) - Kinship Verification Based on Cross-Generation Feature Interaction
Learning [53.62256887837659]
Kinship verification from facial images has been recognized as an emerging yet challenging technique in computer vision applications.
We propose a novel cross-generation feature interaction learning (CFIL) framework for robust kinship verification.
arXiv Detail & Related papers (2021-09-07T01:50:50Z) - Interpretable Multi-dataset Evaluation for Named Entity Recognition [110.64368106131062]
We present a general methodology for interpretable evaluation for the named entity recognition (NER) task.
The proposed evaluation method enables us to interpret the differences in models and datasets, as well as the interplay between them.
By making our analysis tool available, we make it easy for future researchers to run similar analyses and drive progress in this area.
arXiv Detail & Related papers (2020-11-13T10:53:27Z) - On the Relationship Between Active Inference and Control as Inference [62.997667081978825]
Active Inference (AIF) is an emerging framework in the brain sciences which suggests that biological agents act to minimise a variational bound on model evidence.
Control-as-Inference (CAI) is a framework within reinforcement learning which casts decision making as a variational inference problem.
arXiv Detail & Related papers (2020-06-23T13:03:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.