Bayesian Importance of Features (BIF)
- URL: http://arxiv.org/abs/2010.13872v2
- Date: Sat, 17 Sep 2022 22:17:54 GMT
- Title: Bayesian Importance of Features (BIF)
- Authors: Kamil Adamczewski, Frederik Harder, Mijung Park
- Abstract summary: We use the Dirichlet distribution to define the importance of input features and learn it via approximate Bayesian inference.
The learned importance has probabilistic interpretation and provides the relative significance of each input feature to a model's output.
We show the effectiveness of our method on a variety of synthetic and real datasets.
- Score: 11.312036995195594
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We introduce a simple and intuitive framework that provides quantitative
explanations of statistical models through the probabilistic assessment of
input feature importance. The core idea comes from utilizing the Dirichlet
distribution to define the importance of input features and learning it via
approximate Bayesian inference. The learned importance has probabilistic
interpretation and provides the relative significance of each input feature to
a model's output, additionally assessing confidence about its importance
quantification. As a consequence of using the Dirichlet distribution over the
explanations, we can define a closed-form divergence to gauge the similarity
between learned importance under different models. We use this divergence to
study the feature importance explainability tradeoffs with essential notions in
modern machine learning, such as privacy and fairness. Furthermore, BIF can
work on two levels: global explanation (feature importance across all data
instances) and local explanation (individual feature importance for each data
instance). We show the effectiveness of our method on a variety of synthetic
and real datasets, taking into account both tabular and image datasets. The
code is available at https://github.com/kamadforge/featimp_dp.
Related papers
- Sufficient and Necessary Explanations (and What Lies in Between) [6.9035001722324685]
We study two precise notions of feature importance for general machine learning models: sufficiency and necessity.
We propose a unified notion of importance that circumvents these limitations by exploring a continuum along a necessity-sufficiency axis.
arXiv Detail & Related papers (2024-09-30T15:50:57Z) - Gaussian Mixture Models for Affordance Learning using Bayesian Networks [50.18477618198277]
Affordances are fundamental descriptors of relationships between actions, objects and effects.
This paper approaches the problem of an embodied agent exploring the world and learning these affordances autonomously from its sensory experiences.
arXiv Detail & Related papers (2024-02-08T22:05:45Z) - On the Joint Interaction of Models, Data, and Features [82.60073661644435]
We introduce a new tool, the interaction tensor, for empirically analyzing the interaction between data and model through features.
Based on these observations, we propose a conceptual framework for feature learning.
Under this framework, the expected accuracy for a single hypothesis and agreement for a pair of hypotheses can both be derived in closed-form.
arXiv Detail & Related papers (2023-06-07T21:35:26Z) - Exploring the Trade-off between Plausibility, Change Intensity and
Adversarial Power in Counterfactual Explanations using Multi-objective
Optimization [73.89239820192894]
We argue that automated counterfactual generation should regard several aspects of the produced adversarial instances.
We present a novel framework for the generation of counterfactual examples.
arXiv Detail & Related papers (2022-05-20T15:02:53Z) - Interpretable Multi-dataset Evaluation for Named Entity Recognition [110.64368106131062]
We present a general methodology for interpretable evaluation for the named entity recognition (NER) task.
The proposed evaluation method enables us to interpret the differences in models and datasets, as well as the interplay between them.
By making our analysis tool available, we make it easy for future researchers to run similar analyses and drive progress in this area.
arXiv Detail & Related papers (2020-11-13T10:53:27Z) - Towards Unifying Feature Attribution and Counterfactual Explanations:
Different Means to the Same End [17.226134854746267]
We present a method to generate feature attribution explanations from a set of counterfactual examples.
We show how counterfactual examples can be used to evaluate the goodness of an attribution-based explanation in terms of its necessity and sufficiency.
arXiv Detail & Related papers (2020-11-10T05:41:43Z) - Towards a More Reliable Interpretation of Machine Learning Outputs for
Safety-Critical Systems using Feature Importance Fusion [0.0]
We introduce a novel fusion metric and compare it to the state-of-the-art.
Our approach is tested on synthetic data, where the ground truth is known.
Results show that our feature importance ensemble Framework overall produces 15% less feature importance error compared to existing methods.
arXiv Detail & Related papers (2020-09-11T15:51:52Z) - Focus of Attention Improves Information Transfer in Visual Features [80.22965663534556]
This paper focuses on unsupervised learning for transferring visual information in a truly online setting.
The computation of the entropy terms is carried out by a temporal process which yields online estimation of the entropy terms.
In order to better structure the input probability distribution, we use a human-like focus of attention model.
arXiv Detail & Related papers (2020-06-16T15:07:25Z) - Adversarial Infidelity Learning for Model Interpretation [43.37354056251584]
We propose a Model-agnostic Effective Efficient Direct (MEED) IFS framework for model interpretation.
Our framework mitigates concerns about sanity, shortcuts, model identifiability, and information transmission.
Our AIL mechanism can help learn the desired conditional distribution between selected features and targets.
arXiv Detail & Related papers (2020-06-09T16:27:17Z) - Explaining Black Box Predictions and Unveiling Data Artifacts through
Influence Functions [55.660255727031725]
Influence functions explain the decisions of a model by identifying influential training examples.
We conduct a comparison between influence functions and common word-saliency methods on representative tasks.
We develop a new measure based on influence functions that can reveal artifacts in training data.
arXiv Detail & Related papers (2020-05-14T00:45:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.