Marginal Contribution Feature Importance -- an Axiomatic Approach for
The Natural Case
- URL: http://arxiv.org/abs/2010.07910v1
- Date: Thu, 15 Oct 2020 17:41:42 GMT
- Title: Marginal Contribution Feature Importance -- an Axiomatic Approach for
The Natural Case
- Authors: Amnon Catav, Boyang Fu, Jason Ernst, Sriram Sankararaman, Ran
Gilad-Bachrach
- Abstract summary: We develop a set of axioms that represent the properties expected from a feature importance function in the natural scenario.
We analyze this function for its theoretical and empirical properties and compare it to other feature importance scores.
- Score: 13.94406695921022
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: When training a predictive model over medical data, the goal is sometimes to
gain insights about a certain disease. In such cases, it is common to use
feature importance as a tool to highlight significant factors contributing to
that disease. As there are many existing methods for computing feature
importance scores, understanding their relative merits is not trivial. Further,
the diversity of scenarios in which they are used lead to different
expectations from the feature importance scores. While it is common to make the
distinction between local scores that focus on individual predictions and
global scores that look at the contribution of a feature to the model, another
important division distinguishes model scenarios, in which the goal is to
understand predictions of a given model from natural scenarios, in which the
goal is to understand a phenomenon such as a disease. We develop a set of
axioms that represent the properties expected from a feature importance
function in the natural scenario and prove that there exists only one function
that satisfies all of them, the Marginal Contribution Feature Importance (MCI).
We analyze this function for its theoretical and empirical properties and
compare it to other feature importance scores. While our focus is the natural
scenario, we suggest that our axiomatic approach could be carried out in other
scenarios too.
Related papers
- Identifiable Latent Neural Causal Models [82.14087963690561]
Causal representation learning seeks to uncover latent, high-level causal representations from low-level observed data.
We determine the types of distribution shifts that do contribute to the identifiability of causal representations.
We translate our findings into a practical algorithm, allowing for the acquisition of reliable latent causal representations.
arXiv Detail & Related papers (2024-03-23T04:13:55Z) - Detection and Evaluation of bias-inducing Features in Machine learning [14.045499740240823]
In the context of machine learning (ML), one can use cause-to-effect analysis to understand the reason for the biased behavior of the system.
We propose an approach for systematically identifying all bias-inducing features of a model to help support the decision-making of domain experts.
arXiv Detail & Related papers (2023-10-19T15:01:16Z) - On the Joint Interaction of Models, Data, and Features [82.60073661644435]
We introduce a new tool, the interaction tensor, for empirically analyzing the interaction between data and model through features.
Based on these observations, we propose a conceptual framework for feature learning.
Under this framework, the expected accuracy for a single hypothesis and agreement for a pair of hypotheses can both be derived in closed-form.
arXiv Detail & Related papers (2023-06-07T21:35:26Z) - Nonparametric Identifiability of Causal Representations from Unknown
Interventions [63.1354734978244]
We study causal representation learning, the task of inferring latent causal variables and their causal relations from mixtures of the variables.
Our goal is to identify both the ground truth latents and their causal graph up to a set of ambiguities which we show to be irresolvable from interventional data.
arXiv Detail & Related papers (2023-06-01T10:51:58Z) - Inherent Inconsistencies of Feature Importance [6.02357145653815]
Feature importance is a method that assigns scores to the contribution of individual features on prediction outcomes.
This paper presents an axiomatic framework designed to establish coherent relationships among the different contexts of feature importance scores.
arXiv Detail & Related papers (2022-06-16T14:21:51Z) - Masked prediction tasks: a parameter identifiability view [49.533046139235466]
We focus on the widely used self-supervised learning method of predicting masked tokens.
We show that there is a rich landscape of possibilities, out of which some prediction tasks yield identifiability, while others do not.
arXiv Detail & Related papers (2022-02-18T17:09:32Z) - Empirical Estimates on Hand Manipulation are Recoverable: A Step Towards
Individualized and Explainable Robotic Support in Everyday Activities [80.37857025201036]
Key challenge for robotic systems is to figure out the behavior of another agent.
Processing correct inferences is especially challenging when (confounding) factors are not controlled experimentally.
We propose equipping robots with the necessary tools to conduct observational studies on people.
arXiv Detail & Related papers (2022-01-27T22:15:56Z) - On Shapley Credit Allocation for Interpretability [1.52292571922932]
We emphasize the importance of asking the right question when interpreting the decisions of a learning model.
This paper quantifies feature relevance by weaving different natures of interpretations together with different measures as characteristic functions for Shapley symmetrization.
arXiv Detail & Related papers (2020-12-10T08:25:32Z) - Estimating Structural Target Functions using Machine Learning and
Influence Functions [103.47897241856603]
We propose a new framework for statistical machine learning of target functions arising as identifiable functionals from statistical models.
This framework is problem- and model-agnostic and can be used to estimate a broad variety of target parameters of interest in applied statistics.
We put particular focus on so-called coarsening at random/doubly robust problems with partially unobserved information.
arXiv Detail & Related papers (2020-08-14T16:48:29Z) - Nonparametric Feature Impact and Importance [0.6123324869194193]
We give mathematical definitions of feature impact and importance, derived from partial dependence curves, that operate directly on the data.
To assess quality, we show that features ranked by these definitions are competitive with existing feature selection techniques.
arXiv Detail & Related papers (2020-06-08T17:07:35Z) - Understanding Global Feature Contributions With Additive Importance
Measures [14.50261153230204]
We explore the perspective of defining feature importance through the predictive power associated with each feature.
We introduce two notions of predictive power (model-based and universal) and formalize this approach with a framework of additive importance measures.
We then propose SAGE, a model-agnostic method that quantifies predictive power while accounting for feature interactions.
arXiv Detail & Related papers (2020-04-01T19:17:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.