Locally Interpretable Model Agnostic Explanations using Gaussian
Processes
- URL: http://arxiv.org/abs/2108.06907v1
- Date: Mon, 16 Aug 2021 05:49:01 GMT
- Title: Locally Interpretable Model Agnostic Explanations using Gaussian
Processes
- Authors: Aditya Saini, Ranjitha Prasad
- Abstract summary: Local Interpretable Model-Agnostic Explanations (LIME) is a popular technique for explaining the prediction of a single instance.
We propose a Gaussian Process (GP) based variation of locally interpretable models.
We demonstrate that the proposed technique is able to generate faithful explanations using much fewer samples as compared to LIME.
- Score: 2.9189409618561966
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Owing to tremendous performance improvements in data-intensive domains,
machine learning (ML) has garnered immense interest in the research community.
However, these ML models turn out to be black boxes, which are tough to
interpret, resulting in a direct decrease in productivity. Local Interpretable
Model-Agnostic Explanations (LIME) is a popular technique for explaining the
prediction of a single instance. Although LIME is simple and versatile, it
suffers from instability in the generated explanations. In this paper, we
propose a Gaussian Process (GP) based variation of locally interpretable
models. We employ a smart sampling strategy based on the acquisition functions
in Bayesian optimization. Further, we employ the automatic relevance
determination based covariance function in GP, with separate length-scale
parameters for each feature, where the reciprocal of lengthscale parameters
serve as feature explanations. We illustrate the performance of the proposed
technique on two real-world datasets, and demonstrate the superior stability of
the proposed technique. Furthermore, we demonstrate that the proposed technique
is able to generate faithful explanations using much fewer samples as compared
to LIME.
Related papers
- Polynomial Chaos Expanded Gaussian Process [2.287415292857564]
In complex and unknown processes, global models are initially generated over the entire experimental space.
This study addresses the need for models that effectively represent both global and local experimental spaces.
arXiv Detail & Related papers (2024-05-02T07:11:05Z) - Online Variational Sequential Monte Carlo [49.97673761305336]
We build upon the variational sequential Monte Carlo (VSMC) method, which provides computationally efficient and accurate model parameter estimation and Bayesian latent-state inference.
Online VSMC is capable of performing efficiently, entirely on-the-fly, both parameter estimation and particle proposal adaptation.
arXiv Detail & Related papers (2023-12-19T21:45:38Z) - Efficient Model-Free Exploration in Low-Rank MDPs [76.87340323826945]
Low-Rank Markov Decision Processes offer a simple, yet expressive framework for RL with function approximation.
Existing algorithms are either (1) computationally intractable, or (2) reliant upon restrictive statistical assumptions.
We propose the first provably sample-efficient algorithm for exploration in Low-Rank MDPs.
arXiv Detail & Related papers (2023-07-08T15:41:48Z) - Optimizing Hyperparameters with Conformal Quantile Regression [7.316604052864345]
We propose to leverage conformalized quantile regression which makes minimal assumptions about the observation noise.
This translates to quicker HPO convergence on empirical benchmarks.
arXiv Detail & Related papers (2023-05-05T15:33:39Z) - Optimization of Annealed Importance Sampling Hyperparameters [77.34726150561087]
Annealed Importance Sampling (AIS) is a popular algorithm used to estimates the intractable marginal likelihood of deep generative models.
We present a parameteric AIS process with flexible intermediary distributions and optimize the bridging distributions to use fewer number of steps for sampling.
We assess the performance of our optimized AIS for marginal likelihood estimation of deep generative models and compare it to other estimators.
arXiv Detail & Related papers (2022-09-27T07:58:25Z) - An Additive Instance-Wise Approach to Multi-class Model Interpretation [53.87578024052922]
Interpretable machine learning offers insights into what factors drive a certain prediction of a black-box system.
Existing methods mainly focus on selecting explanatory input features, which follow either locally additive or instance-wise approaches.
This work exploits the strengths of both methods and proposes a global framework for learning local explanations simultaneously for multiple target classes.
arXiv Detail & Related papers (2022-07-07T06:50:27Z) - MACE: An Efficient Model-Agnostic Framework for Counterfactual
Explanation [132.77005365032468]
We propose a novel framework of Model-Agnostic Counterfactual Explanation (MACE)
In our MACE approach, we propose a novel RL-based method for finding good counterfactual examples and a gradient-less descent method for improving proximity.
Experiments on public datasets validate the effectiveness with better validity, sparsity and proximity.
arXiv Detail & Related papers (2022-05-31T04:57:06Z) - Surrogate modeling for Bayesian optimization beyond a single Gaussian
process [62.294228304646516]
We propose a novel Bayesian surrogate model to balance exploration with exploitation of the search space.
To endow function sampling with scalability, random feature-based kernel approximation is leveraged per GP model.
To further establish convergence of the proposed EGP-TS to the global optimum, analysis is conducted based on the notion of Bayesian regret.
arXiv Detail & Related papers (2022-05-27T16:43:10Z) - Generalised Gaussian Process Latent Variable Models (GPLVM) with
Stochastic Variational Inference [9.468270453795409]
We study the doubly formulation of the BayesianVM model amenable with minibatch training.
We show how this framework is compatible with different latent variable formulations and perform experiments to compare a suite of models.
We demonstrate how we can train in the presence of massively missing data and obtain high-fidelity reconstructions.
arXiv Detail & Related papers (2022-02-25T21:21:51Z) - Gaussian Process Regression with Local Explanation [28.90948136731314]
We propose GPR with local explanation, which reveals the feature contributions to the prediction of each sample.
In the proposed model, both the prediction and explanation for each sample are performed using an easy-to-interpret locally linear model.
For a new test sample, the proposed model can predict the values of its target variable and weight vector, as well as their uncertainties.
arXiv Detail & Related papers (2020-07-03T13:22:24Z) - A Modified Perturbed Sampling Method for Local Interpretable
Model-agnostic Explanation [35.281127405430674]
Local Interpretable Model-agnostic Explanation (LIME) is a technique that explains the predictions of any classifier faithfully.
This paper proposes a novel Modified Perturbed Sampling operation for LIME (MPS-LIME)
In image classification, MPS-LIME converts the superpixel image into an undirected graph.
arXiv Detail & Related papers (2020-02-18T09:03:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.