Mechanistic Interpretation of Machine Learning Inference: A Fuzzy
Feature Importance Fusion Approach
- URL: http://arxiv.org/abs/2110.11713v1
- Date: Fri, 22 Oct 2021 11:22:21 GMT
- Title: Mechanistic Interpretation of Machine Learning Inference: A Fuzzy
Feature Importance Fusion Approach
- Authors: Divish Rengasamy, Jimiama M. Mase, Mercedes Torres Torres, Benjamin
Rothwell, David A. Winkler, Grazziela P. Figueredo
- Abstract summary: There is a lack of consensus regarding how feature importance should be quantified.
Current state-of-the-art ensemble feature importance fusion uses crisp techniques to fuse results from different approaches.
Here we show how the use of fuzzy data fusion methods can overcome some of the important limitations of crisp fusion methods.
- Score: 0.39146761527401425
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: With the widespread use of machine learning to support decision-making, it is
increasingly important to verify and understand the reasons why a particular
output is produced. Although post-training feature importance approaches assist
this interpretation, there is an overall lack of consensus regarding how
feature importance should be quantified, making explanations of model
predictions unreliable. In addition, many of these explanations depend on the
specific machine learning approach employed and on the subset of data used when
calculating feature importance. A possible solution to improve the reliability
of explanations is to combine results from multiple feature importance
quantifiers from different machine learning approaches coupled with
re-sampling. Current state-of-the-art ensemble feature importance fusion uses
crisp techniques to fuse results from different approaches. There is, however,
significant loss of information as these approaches are not context-aware and
reduce several quantifiers to a single crisp output. More importantly, their
representation of 'importance' as coefficients is misleading and
incomprehensible to end-users and decision makers. Here we show how the use of
fuzzy data fusion methods can overcome some of the important limitations of
crisp fusion methods.
Related papers
- A Critical Assessment of Interpretable and Explainable Machine Learning for Intrusion Detection [0.0]
We study the use of overly complex and opaque ML models, unaccounted data imbalances and correlated features, inconsistent influential features across different explanation methods, and the implausible utility of explanations.
Specifically, we advise avoiding complex opaque models such as Deep Neural Networks and instead using interpretable ML models such as Decision Trees.
We find that feature-based model explanations are most often inconsistent across different settings.
arXiv Detail & Related papers (2024-07-04T15:35:42Z) - Matched Machine Learning: A Generalized Framework for Treatment Effect
Inference With Learned Metrics [87.05961347040237]
We introduce Matched Machine Learning, a framework that combines the flexibility of machine learning black boxes with the interpretability of matching.
Our framework uses machine learning to learn an optimal metric for matching units and estimating outcomes.
We show empirically that instances of Matched Machine Learning perform on par with black-box machine learning methods and better than existing matching methods for similar problems.
arXiv Detail & Related papers (2023-04-03T19:32:30Z) - Explainable Data-Driven Optimization: From Context to Decision and Back
Again [76.84947521482631]
Data-driven optimization uses contextual information and machine learning algorithms to find solutions to decision problems with uncertain parameters.
We introduce a counterfactual explanation methodology tailored to explain solutions to data-driven problems.
We demonstrate our approach by explaining key problems in operations management such as inventory management and routing.
arXiv Detail & Related papers (2023-01-24T15:25:16Z) - Interpretability with full complexity by constraining feature
information [1.52292571922932]
Interpretability is a pressing issue for machine learning.
We approach interpretability from a new angle: constrain the information about the features without restricting the complexity of the model.
We develop a framework for extracting insight from the spectrum of approximate models.
arXiv Detail & Related papers (2022-11-30T18:59:01Z) - EFI: A Toolbox for Feature Importance Fusion and Interpretation in
Python [1.593222804814135]
Ensemble Feature Importance (EFI) is an open-source Python toolbox for machine learning (ML) researchers, domain experts, and decision makers.
EFI provides robust and accurate feature importance quantification and more reliable mechanistic interpretation of feature importance for prediction problems.
arXiv Detail & Related papers (2022-08-08T18:02:37Z) - BayesIMP: Uncertainty Quantification for Causal Data Fusion [52.184885680729224]
We study the causal data fusion problem, where datasets pertaining to multiple causal graphs are combined to estimate the average treatment effect of a target variable.
We introduce a framework which combines ideas from probabilistic integration and kernel mean embeddings to represent interventional distributions in the reproducing kernel Hilbert space.
arXiv Detail & Related papers (2021-06-07T10:14:18Z) - Beyond Trivial Counterfactual Explanations with Diverse Valuable
Explanations [64.85696493596821]
In computer vision applications, generative counterfactual methods indicate how to perturb a model's input to change its prediction.
We propose a counterfactual method that learns a perturbation in a disentangled latent space that is constrained using a diversity-enforcing loss.
Our model improves the success rate of producing high-quality valuable explanations when compared to previous state-of-the-art methods.
arXiv Detail & Related papers (2021-03-18T12:57:34Z) - Bayesian Importance of Features (BIF) [11.312036995195594]
We use the Dirichlet distribution to define the importance of input features and learn it via approximate Bayesian inference.
The learned importance has probabilistic interpretation and provides the relative significance of each input feature to a model's output.
We show the effectiveness of our method on a variety of synthetic and real datasets.
arXiv Detail & Related papers (2020-10-26T19:55:58Z) - Accurate and Robust Feature Importance Estimation under Distribution
Shifts [49.58991359544005]
PRoFILE is a novel feature importance estimation method.
We show significant improvements over state-of-the-art approaches, both in terms of fidelity and robustness.
arXiv Detail & Related papers (2020-09-30T05:29:01Z) - Towards a More Reliable Interpretation of Machine Learning Outputs for
Safety-Critical Systems using Feature Importance Fusion [0.0]
We introduce a novel fusion metric and compare it to the state-of-the-art.
Our approach is tested on synthetic data, where the ground truth is known.
Results show that our feature importance ensemble Framework overall produces 15% less feature importance error compared to existing methods.
arXiv Detail & Related papers (2020-09-11T15:51:52Z) - Fairness by Learning Orthogonal Disentangled Representations [50.82638766862974]
We propose a novel disentanglement approach to invariant representation problem.
We enforce the meaningful representation to be agnostic to sensitive information by entropy.
The proposed approach is evaluated on five publicly available datasets.
arXiv Detail & Related papers (2020-03-12T11:09:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.