Shapley variable importance clouds for interpretable machine learning
- URL: http://arxiv.org/abs/2110.02484v1
- Date: Wed, 6 Oct 2021 03:41:04 GMT
- Title: Shapley variable importance clouds for interpretable machine learning
- Authors: Yilin Ning, Marcus Eng Hock Ong, Bibhas Chakraborty, Benjamin Alan
Goldstein, Daniel Shu Wei Ting, Roger Vaughan, Nan Liu
- Abstract summary: We propose a Shapley variable importance cloud that pools information across good models to avoid biased assessments in SHAP analyses of final models.
We demonstrate the additional insights gain compared to conventional explanations and Dong and Rudin's method using criminal justice and electronic medical records data.
- Score: 2.830197032154301
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Interpretable machine learning has been focusing on explaining final models
that optimize performance. The current state-of-the-art is the Shapley additive
explanations (SHAP) that locally explains variable impact on individual
predictions, and it is recently extended for a global assessment across the
dataset. Recently, Dong and Rudin proposed to extend the investigation to
models from the same class as the final model that are "good enough", and
identified a previous overclaim of variable importance based on a single model.
However, this method does not directly integrate with existing Shapley-based
interpretations. We close this gap by proposing a Shapley variable importance
cloud that pools information across good models to avoid biased assessments in
SHAP analyses of final models, and communicate the findings via novel
visualizations. We demonstrate the additional insights gain compared to
conventional explanations and Dong and Rudin's method using criminal justice
and electronic medical records data.
Related papers
- Learning Augmentation Policies from A Model Zoo for Time Series Forecasting [58.66211334969299]
We introduce AutoTSAug, a learnable data augmentation method based on reinforcement learning.
By augmenting the marginal samples with a learnable policy, AutoTSAug substantially improves forecasting performance.
arXiv Detail & Related papers (2024-09-10T07:34:19Z) - Shapley Marginal Surplus for Strong Models [0.9831489366502301]
We show that while Shapley values might be accurate explainers of model predictions, machine learning models themselves are often poor explainers of the true data-generating process (DGP)
We introduce a novel variable importance algorithm, Shapley Marginal Surplus for Strong Models, that samples the space of possible models to come up with an inferential measure of feature importance.
arXiv Detail & Related papers (2024-08-16T17:06:07Z) - Data Shapley in One Training Run [88.59484417202454]
Data Shapley provides a principled framework for attributing data's contribution within machine learning contexts.
Existing approaches require re-training models on different data subsets, which is computationally intensive.
This paper introduces In-Run Data Shapley, which addresses these limitations by offering scalable data attribution for a target model of interest.
arXiv Detail & Related papers (2024-06-16T17:09:24Z) - Predictive Churn with the Set of Good Models [64.05949860750235]
We study the effect of conflicting predictions over the set of near-optimal machine learning models.
We present theoretical results on the expected churn between models within the Rashomon set.
We show how our approach can be used to better anticipate, reduce, and avoid churn in consumer-facing applications.
arXiv Detail & Related papers (2024-02-12T16:15:25Z) - Shapley variable importance cloud for machine learning models [4.1359299555083595]
Recently developed Shapley variable importance cloud (ShapleyVIC) provides comprehensive and robust variable importance assessments.
benefits of ShapleyVIC inference have been demonstrated in real-life prediction tasks.
ShapleyVIC implementation for machine learning models to enable wider applications.
arXiv Detail & Related papers (2022-12-16T09:45:22Z) - Consistent Counterfactuals for Deep Models [25.1271020453651]
Counterfactual examples are used to explain predictions of machine learning models in key areas such as finance and medical diagnosis.
This paper studies the consistency of model prediction on counterfactual examples in deep networks under small changes to initial training conditions.
arXiv Detail & Related papers (2021-10-06T23:48:55Z) - Back2Future: Leveraging Backfill Dynamics for Improving Real-time
Predictions in Future [73.03458424369657]
In real-time forecasting in public health, data collection is a non-trivial and demanding task.
'Backfill' phenomenon and its effect on model performance has been barely studied in the prior literature.
We formulate a novel problem and neural framework Back2Future that aims to refine a given model's predictions in real-time.
arXiv Detail & Related papers (2021-06-08T14:48:20Z) - EDDA: Explanation-driven Data Augmentation to Improve Model and
Explanation Alignment [12.729179495550557]
We seek a methodology that can improve alignment between model predictions and explanation method.
We achieve this through a novel explanation-driven data augmentation (EDDA) method.
This is based on the simple motivating principle that occluding salient regions for the model prediction should decrease the model confidence in the prediction.
arXiv Detail & Related papers (2021-05-29T00:42:42Z) - Analysis and modeling to forecast in time series: a systematic review [0.0]
This paper surveys state-of-the-art methods and models dedicated to time series analysis and modeling, with the final aim of prediction.
This review aims to offer a structured and comprehensive view of the full process flow, and encompasses time series decomposition, stationary tests, modeling and forecasting.
arXiv Detail & Related papers (2021-03-31T23:48:46Z) - Beyond Trivial Counterfactual Explanations with Diverse Valuable
Explanations [64.85696493596821]
In computer vision applications, generative counterfactual methods indicate how to perturb a model's input to change its prediction.
We propose a counterfactual method that learns a perturbation in a disentangled latent space that is constrained using a diversity-enforcing loss.
Our model improves the success rate of producing high-quality valuable explanations when compared to previous state-of-the-art methods.
arXiv Detail & Related papers (2021-03-18T12:57:34Z) - Generative Temporal Difference Learning for Infinite-Horizon Prediction [101.59882753763888]
We introduce the $gamma$-model, a predictive model of environment dynamics with an infinite probabilistic horizon.
We discuss how its training reflects an inescapable tradeoff between training-time and testing-time compounding errors.
arXiv Detail & Related papers (2020-10-27T17:54:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.