Trade-off Between Efficiency and Consistency for Removal-based
Explanations
- URL: http://arxiv.org/abs/2210.17426v3
- Date: Fri, 20 Oct 2023 14:42:34 GMT
- Title: Trade-off Between Efficiency and Consistency for Removal-based
Explanations
- Authors: Yifan Zhang, Haowei He, Zhiquan Tan, Yang Yuan
- Abstract summary: We establish the Impossible Trinity Theorem, which posits that interpretability, efficiency, and consistency cannot hold simultaneously.
Recognizing that the attainment of an ideal explanation remains elusive, we propose the utilization of interpretation error as a metric to gauge inefficiencies and inconsistencies.
Our empirical findings indicate that the proposed methods achieve a substantial reduction in interpretation error, up to 31.8 times lower when compared to alternative techniques.
- Score: 12.338207007436566
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the current landscape of explanation methodologies, most predominant
approaches, such as SHAP and LIME, employ removal-based techniques to evaluate
the impact of individual features by simulating various scenarios with specific
features omitted. Nonetheless, these methods primarily emphasize efficiency in
the original context, often resulting in general inconsistencies. In this
paper, we demonstrate that such inconsistency is an inherent aspect of these
approaches by establishing the Impossible Trinity Theorem, which posits that
interpretability, efficiency, and consistency cannot hold simultaneously.
Recognizing that the attainment of an ideal explanation remains elusive, we
propose the utilization of interpretation error as a metric to gauge
inefficiencies and inconsistencies. To this end, we present two novel
algorithms founded on the standard polynomial basis, aimed at minimizing
interpretation error. Our empirical findings indicate that the proposed methods
achieve a substantial reduction in interpretation error, up to 31.8 times lower
when compared to alternative techniques. Code is available at
https://github.com/trusty-ai/efficient-consistent-explanations.
Related papers
- Exogenous Matching: Learning Good Proposals for Tractable Counterfactual Estimation [1.9662978733004601]
We propose an importance sampling method for tractable and efficient estimation of counterfactual expressions.
By minimizing a common upper bound of counterfactual estimators, we transform the variance minimization problem into a conditional distribution learning problem.
We validate the theoretical results through experiments under various types and settings of Structural Causal Models (SCMs) and demonstrate the outperformance on counterfactual estimation tasks.
arXiv Detail & Related papers (2024-10-17T03:08:28Z) - Algorithmic Recourse with Missing Values [11.401006371457436]
This paper proposes a new framework of algorithmic recourse (AR) that works even in the presence of missing values.
AR aims to provide a recourse action for altering the undesired prediction result given by a classifier.
Experimental results demonstrated the efficacy of our method in the presence of missing values compared to the baselines.
arXiv Detail & Related papers (2023-04-28T03:22:48Z) - Synergies between Disentanglement and Sparsity: Generalization and
Identifiability in Multi-Task Learning [79.83792914684985]
We prove a new identifiability result that provides conditions under which maximally sparse base-predictors yield disentangled representations.
Motivated by this theoretical result, we propose a practical approach to learn disentangled representations based on a sparsity-promoting bi-level optimization problem.
arXiv Detail & Related papers (2022-11-26T21:02:09Z) - Making Linear MDPs Practical via Contrastive Representation Learning [101.75885788118131]
It is common to address the curse of dimensionality in Markov decision processes (MDPs) by exploiting low-rank representations.
We consider an alternative definition of linear MDPs that automatically ensures normalization while allowing efficient representation learning.
We demonstrate superior performance over existing state-of-the-art model-based and model-free algorithms on several benchmarks.
arXiv Detail & Related papers (2022-07-14T18:18:02Z) - Self-attention fusion for audiovisual emotion recognition with
incomplete data [103.70855797025689]
We consider the problem of multimodal data analysis with a use case of audiovisual emotion recognition.
We propose an architecture capable of learning from raw data and describe three variants of it with distinct modality fusion mechanisms.
arXiv Detail & Related papers (2022-01-26T18:04:29Z) - Regularizing Variational Autoencoder with Diversity and Uncertainty
Awareness [61.827054365139645]
Variational Autoencoder (VAE) approximates the posterior of latent variables based on amortized variational inference.
We propose an alternative model, DU-VAE, for learning a more Diverse and less Uncertain latent space.
arXiv Detail & Related papers (2021-10-24T07:58:13Z) - Discriminative Attribution from Counterfactuals [64.94009515033984]
We present a method for neural network interpretability by combining feature attribution with counterfactual explanations.
We show that this method can be used to quantitatively evaluate the performance of feature attribution methods in an objective manner.
arXiv Detail & Related papers (2021-09-28T00:53:34Z) - On Sample Based Explanation Methods for NLP:Efficiency, Faithfulness,
and Semantic Evaluation [23.72825603188359]
We can improve the interpretability of explanations by allowing arbitrary text sequences as the explanation unit.
We propose a semantic-based evaluation metric that can better align with humans' judgment of explanations.
arXiv Detail & Related papers (2021-06-09T00:49:56Z) - On Guaranteed Optimal Robust Explanations for NLP Models [16.358394218953833]
We build on abduction-based explanations for ma-chine learning and develop a method for computing local explanations for neural network models.
We present two solution algorithms, respectively based on implicit hitting sets and maximum universal subsets.
We evaluate our framework on three widely used sentiment analysis tasks and texts of up to100words from SST, Twitter and IMDB datasets.
arXiv Detail & Related papers (2021-05-08T08:44:48Z) - Learning explanations that are hard to vary [75.30552491694066]
We show that averaging across examples can favor memorization and patchwork' solutions that sew together different strategies.
We then propose and experimentally validate a simple alternative algorithm based on a logical AND.
arXiv Detail & Related papers (2020-09-01T10:17:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.