GLIME: General, Stable and Local LIME Explanation
- URL: http://arxiv.org/abs/2311.15722v1
- Date: Mon, 27 Nov 2023 11:17:20 GMT
- Title: GLIME: General, Stable and Local LIME Explanation
- Authors: Zeren Tan, Yang Tian, Jian Li
- Abstract summary: Local Interpretable Model-agnostic Explanations (LIME) is a widely adpoted method for understanding model behaviors.
We introduce GLIME, an enhanced framework extending LIME and unifying several prior methods.
By employing a local and unbiased sampling distribution, GLIME generates explanations with higher local fidelity compared to LIME.
- Score: 11.002828804775392
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As black-box machine learning models grow in complexity and find applications
in high-stakes scenarios, it is imperative to provide explanations for their
predictions. Although Local Interpretable Model-agnostic Explanations (LIME)
[22] is a widely adpoted method for understanding model behaviors, it is
unstable with respect to random seeds [35,24,3] and exhibits low local fidelity
(i.e., how well the explanation approximates the model's local behaviors)
[21,16]. Our study shows that this instability problem stems from small sample
weights, leading to the dominance of regularization and slow convergence.
Additionally, LIME's sampling neighborhood is non-local and biased towards the
reference, resulting in poor local fidelity and sensitivity to reference
choice. To tackle these challenges, we introduce GLIME, an enhanced framework
extending LIME and unifying several prior methods. Within the GLIME framework,
we derive an equivalent formulation of LIME that achieves significantly faster
convergence and improved stability. By employing a local and unbiased sampling
distribution, GLIME generates explanations with higher local fidelity compared
to LIME. GLIME explanations are independent of reference choice. Moreover,
GLIME offers users the flexibility to choose a sampling distribution based on
their specific scenarios.
Related papers
- MASALA: Model-Agnostic Surrogate Explanations by Locality Adaptation [3.587367153279351]
Existing local Explainable AI (XAI) methods select a region of the input space in the vicinity of a given input instance, for which they approximate the behaviour of a model using a simpler and more interpretable surrogate model.
We propose a novel method, MASALA, for generating explanations, which automatically determines the appropriate local region of impactful model behaviour for each individual instance being explained.
arXiv Detail & Related papers (2024-08-19T15:26:45Z) - Cycles of Thought: Measuring LLM Confidence through Stable Explanations [53.15438489398938]
Large language models (LLMs) can reach and even surpass human-level accuracy on a variety of benchmarks, but their overconfidence in incorrect responses is still a well-documented failure mode.
We propose a framework for measuring an LLM's uncertainty with respect to the distribution of generated explanations for an answer.
arXiv Detail & Related papers (2024-06-05T16:35:30Z) - Guarantee Regions for Local Explanations [29.429229877959663]
We propose an anchor-based algorithm for identifying regions in which local explanations are guaranteed to be correct.
Our method produces an interpretable feature-aligned box where the prediction of the local surrogate model is guaranteed to match the predictive model.
arXiv Detail & Related papers (2024-02-20T06:04:44Z) - Optimal Multi-Distribution Learning [88.3008613028333]
Multi-distribution learning seeks to learn a shared model that minimizes the worst-case risk across $k$ distinct data distributions.
We propose a novel algorithm that yields an varepsilon-optimal randomized hypothesis with a sample complexity on the order of (d+k)/varepsilon2.
arXiv Detail & Related papers (2023-12-08T16:06:29Z) - Numerically assisted determination of local models in network scenarios [55.2480439325792]
We develop a numerical tool for finding explicit local models that reproduce a given statistical behaviour.
We provide conjectures for the critical visibilities of the Greenberger-Horne-Zeilinger (GHZ) and W distributions.
The developed codes and documentation are publicly available at281.com/mariofilho/localmodels.
arXiv Detail & Related papers (2023-03-17T13:24:04Z) - Distributionally Robust Models with Parametric Likelihood Ratios [123.05074253513935]
Three simple ideas allow us to train models with DRO using a broader class of parametric likelihood ratios.
We find that models trained with the resulting parametric adversaries are consistently more robust to subpopulation shifts when compared to other DRO approaches.
arXiv Detail & Related papers (2022-04-13T12:43:12Z) - Locally Invariant Explanations: Towards Stable and Unidirectional
Explanations through Local Invariant Learning [15.886405745163234]
We propose a model agnostic local explanation method inspired by the invariant risk minimization principle.
Our algorithm is simple and efficient to train, and can ascertain stable input features for local decisions of a black-box without access to side information.
arXiv Detail & Related papers (2022-01-28T14:29:25Z) - Locally Interpretable Model Agnostic Explanations using Gaussian
Processes [2.9189409618561966]
Local Interpretable Model-Agnostic Explanations (LIME) is a popular technique for explaining the prediction of a single instance.
We propose a Gaussian Process (GP) based variation of locally interpretable models.
We demonstrate that the proposed technique is able to generate faithful explanations using much fewer samples as compared to LIME.
arXiv Detail & Related papers (2021-08-16T05:49:01Z) - Learning Invariant Representations and Risks for Semi-supervised Domain
Adaptation [109.73983088432364]
We propose the first method that aims to simultaneously learn invariant representations and risks under the setting of semi-supervised domain adaptation (Semi-DA)
We introduce the LIRR algorithm for jointly textbfLearning textbfInvariant textbfRepresentations and textbfRisks.
arXiv Detail & Related papers (2020-10-09T15:42:35Z) - GANs with Variational Entropy Regularizers: Applications in Mitigating
the Mode-Collapse Issue [95.23775347605923]
Building on the success of deep learning, Generative Adversarial Networks (GANs) provide a modern approach to learn a probability distribution from observed samples.
GANs often suffer from the mode collapse issue where the generator fails to capture all existing modes of the input distribution.
We take an information-theoretic approach and maximize a variational lower bound on the entropy of the generated samples to increase their diversity.
arXiv Detail & Related papers (2020-09-24T19:34:37Z) - A Modified Perturbed Sampling Method for Local Interpretable
Model-agnostic Explanation [35.281127405430674]
Local Interpretable Model-agnostic Explanation (LIME) is a technique that explains the predictions of any classifier faithfully.
This paper proposes a novel Modified Perturbed Sampling operation for LIME (MPS-LIME)
In image classification, MPS-LIME converts the superpixel image into an undirected graph.
arXiv Detail & Related papers (2020-02-18T09:03:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.