Using Shapley Values and Variational Autoencoders to Explain Predictive
Models with Dependent Mixed Features
- URL: http://arxiv.org/abs/2111.13507v1
- Date: Fri, 26 Nov 2021 14:05:45 GMT
- Title: Using Shapley Values and Variational Autoencoders to Explain Predictive
Models with Dependent Mixed Features
- Authors: Lars Henry Berge Olsen, Ingrid Kristine Glad, Martin Jullum and
Kjersti Aas
- Abstract summary: We use a variational autoencoder with arbitrary conditioning (VAEAC) to model all feature dependencies simultaneously.
We apply VAEAC to the Abalone data set from the UCI Machine Learning Repository.
- Score: 2.064612766965483
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Shapley values are today extensively used as a model-agnostic explanation
framework to explain complex predictive machine learning models. Shapley values
have desirable theoretical properties and a sound mathematical foundation.
Precise Shapley value estimates for dependent data rely on accurate modeling of
the dependencies between all feature combinations. In this paper, we use a
variational autoencoder with arbitrary conditioning (VAEAC) to model all
feature dependencies simultaneously. We demonstrate through comprehensive
simulation studies that VAEAC outperforms the state-of-the-art methods for a
wide range of settings for both continuous and mixed dependent features.
Finally, we apply VAEAC to the Abalone data set from the UCI Machine Learning
Repository.
Related papers
- Sample Complexity Characterization for Linear Contextual MDPs [67.79455646673762]
Contextual decision processes (CMDPs) describe a class of reinforcement learning problems in which the transition kernels and reward functions can change over time with different MDPs indexed by a context variable.
CMDPs serve as an important framework to model many real-world applications with time-varying environments.
We study CMDPs under two linear function approximation models: Model I with context-varying representations and common linear weights for all contexts; and Model II with common representations for all contexts and context-varying linear weights.
arXiv Detail & Related papers (2024-02-05T03:25:04Z) - Online Variational Sequential Monte Carlo [49.97673761305336]
We build upon the variational sequential Monte Carlo (VSMC) method, which provides computationally efficient and accurate model parameter estimation and Bayesian latent-state inference.
Online VSMC is capable of performing efficiently, entirely on-the-fly, both parameter estimation and particle proposal adaptation.
arXiv Detail & Related papers (2023-12-19T21:45:38Z) - Efficient Shapley Values Estimation by Amortization for Text
Classification [66.7725354593271]
We develop an amortized model that directly predicts each input feature's Shapley Value without additional model evaluations.
Experimental results on two text classification datasets demonstrate that our amortized model estimates Shapley Values accurately with up to 60 times speedup.
arXiv Detail & Related papers (2023-05-31T16:19:13Z) - Shapley variable importance cloud for machine learning models [4.1359299555083595]
Recently developed Shapley variable importance cloud (ShapleyVIC) provides comprehensive and robust variable importance assessments.
benefits of ShapleyVIC inference have been demonstrated in real-life prediction tasks.
ShapleyVIC implementation for machine learning models to enable wider applications.
arXiv Detail & Related papers (2022-12-16T09:45:22Z) - HyperImpute: Generalized Iterative Imputation with Automatic Model
Selection [77.86861638371926]
We propose a generalized iterative imputation framework for adaptively and automatically configuring column-wise models.
We provide a concrete implementation with out-of-the-box learners, simulators, and interfaces.
arXiv Detail & Related papers (2022-06-15T19:10:35Z) - Exact Shapley Values for Local and Model-True Explanations of Decision
Tree Ensembles [0.0]
We consider the application of Shapley values for explaining decision tree ensembles.
We present a novel approach to Shapley value-based feature attribution that can be applied to random forests and boosted decision trees.
arXiv Detail & Related papers (2021-12-16T20:16:02Z) - Counterfactual Shapley Additive Explanations [6.916452769334367]
We propose a variant of SHAP, CoSHAP, that uses counterfactual generation techniques to produce a background dataset.
We motivate the need within the actionable recourse setting for careful consideration of background datasets when using Shapley values for feature attributions.
arXiv Detail & Related papers (2021-10-27T08:44:53Z) - Explaining predictive models using Shapley values and non-parametric
vine copulas [2.6774008509840996]
We propose two new approaches for modelling the dependence between the features.
The performance of the proposed methods is evaluated on simulated data sets and a real data set.
Experiments demonstrate that the vine copula approaches give more accurate approximations to the true Shapley values than its competitors.
arXiv Detail & Related papers (2021-02-12T09:43:28Z) - Anomaly Detection of Time Series with Smoothness-Inducing Sequential
Variational Auto-Encoder [59.69303945834122]
We present a Smoothness-Inducing Sequential Variational Auto-Encoder (SISVAE) model for robust estimation and anomaly detection of time series.
Our model parameterizes mean and variance for each time-stamp with flexible neural networks.
We show the effectiveness of our model on both synthetic datasets and public real-world benchmarks.
arXiv Detail & Related papers (2021-02-02T06:15:15Z) - Autoencoding Variational Autoencoder [56.05008520271406]
We study the implications of this behaviour on the learned representations and also the consequences of fixing it by introducing a notion of self consistency.
We show that encoders trained with our self-consistency approach lead to representations that are robust (insensitive) to perturbations in the input introduced by adversarial attacks.
arXiv Detail & Related papers (2020-12-07T14:16:14Z) - Explaining predictive models with mixed features using Shapley values
and conditional inference trees [1.8065361710947976]
Shapley values stand out as a sound method to explain predictions from any type of machine learning model.
We propose a method to explain mixed dependent features by modeling the dependence structure of the features using conditional inference trees.
arXiv Detail & Related papers (2020-07-02T11:25:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.