Guarantee Regions for Local Explanations
- URL: http://arxiv.org/abs/2402.12737v1
- Date: Tue, 20 Feb 2024 06:04:44 GMT
- Title: Guarantee Regions for Local Explanations
- Authors: Marton Havasi, Sonali Parbhoo, Finale Doshi-Velez
- Abstract summary: We propose an anchor-based algorithm for identifying regions in which local explanations are guaranteed to be correct.
Our method produces an interpretable feature-aligned box where the prediction of the local surrogate model is guaranteed to match the predictive model.
- Score: 29.429229877959663
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Interpretability methods that utilise local surrogate models (e.g. LIME) are
very good at describing the behaviour of the predictive model at a point of
interest, but they are not guaranteed to extrapolate to the local region
surrounding the point. However, overfitting to the local curvature of the
predictive model and malicious tampering can significantly limit extrapolation.
We propose an anchor-based algorithm for identifying regions in which local
explanations are guaranteed to be correct by explicitly describing those
intervals along which the input features can be trusted. Our method produces an
interpretable feature-aligned box where the prediction of the local surrogate
model is guaranteed to match the predictive model. We demonstrate that our
algorithm can be used to find explanations with larger guarantee regions that
better cover the data manifold compared to existing baselines. We also show how
our method can identify misleading local explanations with significantly poorer
guarantee regions.
Related papers
- l_inf-approximation of localized distributions [11.211950982996784]
In this work, we establish a dimension independent error bound for the marginals of approximate distributions.
We show how to use localized likelihood-informed subspace method and localized score matching, which not only avoid dimension dependence in the approximation error, but also significantly reduce the computational cost.
arXiv Detail & Related papers (2024-10-15T16:47:05Z) - MASALA: Model-Agnostic Surrogate Explanations by Locality Adaptation [3.587367153279351]
Existing local Explainable AI (XAI) methods select a region of the input space in the vicinity of a given input instance, for which they approximate the behaviour of a model using a simpler and more interpretable surrogate model.
We propose a novel method, MASALA, for generating explanations, which automatically determines the appropriate local region of impactful model behaviour for each individual instance being explained.
arXiv Detail & Related papers (2024-08-19T15:26:45Z) - GLIME: General, Stable and Local LIME Explanation [11.002828804775392]
Local Interpretable Model-agnostic Explanations (LIME) is a widely adpoted method for understanding model behaviors.
We introduce GLIME, an enhanced framework extending LIME and unifying several prior methods.
By employing a local and unbiased sampling distribution, GLIME generates explanations with higher local fidelity compared to LIME.
arXiv Detail & Related papers (2023-11-27T11:17:20Z) - Semi-Supervised Crowd Counting with Contextual Modeling: Facilitating Holistic Understanding of Crowd Scenes [19.987151025364067]
This paper presents a new semi-supervised method for training a reliable crowd counting model.
We foster the model's intrinsic'subitizing' capability, which allows it to accurately estimate the count in regions.
Our method achieves the state-of-the-art performance, surpassing previous approaches by a large margin on challenging benchmarks.
arXiv Detail & Related papers (2023-10-16T12:42:43Z) - Numerically assisted determination of local models in network scenarios [55.2480439325792]
We develop a numerical tool for finding explicit local models that reproduce a given statistical behaviour.
We provide conjectures for the critical visibilities of the Greenberger-Horne-Zeilinger (GHZ) and W distributions.
The developed codes and documentation are publicly available at281.com/mariofilho/localmodels.
arXiv Detail & Related papers (2023-03-17T13:24:04Z) - RbX: Region-based explanations of prediction models [69.3939291118954]
Region-based explanations (RbX) is a model-agnostic method to generate local explanations of scalar outputs from a black-box prediction model.
RbX is guaranteed to satisfy a "sparsity axiom," which requires that features which do not enter into the prediction model are assigned zero importance.
arXiv Detail & Related papers (2022-10-17T03:38:06Z) - Change Detection for Local Explainability in Evolving Data Streams [72.4816340552763]
Local feature attribution methods have become a popular technique for post-hoc and model-agnostic explanations.
It is often unclear how local attributions behave in realistic, constantly evolving settings such as streaming and online applications.
We present CDLEEDS, a flexible and model-agnostic framework for detecting local change and concept drift.
arXiv Detail & Related papers (2022-09-06T18:38:34Z) - Sampling Based On Natural Image Statistics Improves Local Surrogate
Explainers [111.31448606885672]
Surrogate explainers are a popular post-hoc interpretability method to further understand how a model arrives at a prediction.
We propose two approaches to do so, namely (1) altering the method for sampling the local neighbourhood and (2) using perceptual metrics to convey some of the properties of the distribution of natural images.
arXiv Detail & Related papers (2022-08-08T08:10:13Z) - MD-split+: Practical Local Conformal Inference in High Dimensions [0.5439020425819]
MD-split+ is a practical local conformal approach that creates X partitions based on localized model performance.
We discuss how our local partitions philosophically align with expected behavior from an unattainable conditional conformal inference approach.
arXiv Detail & Related papers (2021-07-07T15:19:16Z) - Scalable Marginal Likelihood Estimation for Model Selection in Deep
Learning [78.83598532168256]
Marginal-likelihood based model-selection is rarely used in deep learning due to estimation difficulties.
Our work shows that marginal likelihoods can improve generalization and be useful when validation data is unavailable.
arXiv Detail & Related papers (2021-04-11T09:50:24Z) - Learning Invariant Representations and Risks for Semi-supervised Domain
Adaptation [109.73983088432364]
We propose the first method that aims to simultaneously learn invariant representations and risks under the setting of semi-supervised domain adaptation (Semi-DA)
We introduce the LIRR algorithm for jointly textbfLearning textbfInvariant textbfRepresentations and textbfRisks.
arXiv Detail & Related papers (2020-10-09T15:42:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.