FASTER-CE: Fast, Sparse, Transparent, and Robust Counterfactual
Explanations
- URL: http://arxiv.org/abs/2210.06578v1
- Date: Wed, 12 Oct 2022 20:41:17 GMT
- Title: FASTER-CE: Fast, Sparse, Transparent, and Robust Counterfactual
Explanations
- Authors: Shubham Sharma, Alan H. Gee, Jette Henderson, Joydeep Ghosh
- Abstract summary: We present FASTER-CE: a novel set of algorithms to generate fast, sparse, and robust counterfactual explanations.
The key idea is to efficiently find promising search directions for counterfactuals in a latent space that is specified via an autoencoder.
The ability to quickly examine combinations of the most promising gradient directions as well as to incorporate additional user-defined constraints allows us to generate multiple counterfactual explanations.
- Score: 9.532938405575788
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Counterfactual explanations have substantially increased in popularity in the
past few years as a useful human-centric way of understanding individual
black-box model predictions. While several properties desired of high-quality
counterfactuals have been identified in the literature, three crucial concerns:
the speed of explanation generation, robustness/sensitivity and succinctness of
explanations (sparsity) have been relatively unexplored. In this paper, we
present FASTER-CE: a novel set of algorithms to generate fast, sparse, and
robust counterfactual explanations. The key idea is to efficiently find
promising search directions for counterfactuals in a latent space that is
specified via an autoencoder. These directions are determined based on
gradients with respect to each of the original input features as well as of the
target, as estimated in the latent space. The ability to quickly examine
combinations of the most promising gradient directions as well as to
incorporate additional user-defined constraints allows us to generate multiple
counterfactual explanations that are sparse, realistic, and robust to input
manipulations. Through experiments on three datasets of varied complexities, we
show that FASTER-CE is not only much faster than other state of the art methods
for generating multiple explanations but also is significantly superior when
considering a larger set of desirable (and often conflicting) properties.
Specifically we present results across multiple performance metrics: sparsity,
proximity, validity, speed of generation, and the robustness of explanations,
to highlight the capabilities of the FASTER-CE family.
Related papers
- Improving Network Interpretability via Explanation Consistency Evaluation [56.14036428778861]
We propose a framework that acquires more explainable activation heatmaps and simultaneously increase the model performance.
Specifically, our framework introduces a new metric, i.e., explanation consistency, to reweight the training samples adaptively in model learning.
Our framework then promotes the model learning by paying closer attention to those training samples with a high difference in explanations.
arXiv Detail & Related papers (2024-08-08T17:20:08Z) - XPrompt:Explaining Large Language Model's Generation via Joint Prompt Attribution [26.639271355209104]
Large Language Models (LLMs) have demonstrated impressive performances in complex text generation tasks.
The contribution of the input prompt to the generated content still remains obscure to humans.
We introduce a counterfactual explanation framework based on joint prompt attribution, XPrompt.
arXiv Detail & Related papers (2024-05-30T18:16:41Z) - Enhancing Counterfactual Explanation Search with Diffusion Distance and Directional Coherence [0.0]
A pressing issue in the adoption of AI models is the increasing demand for more human-centric explanations of their predictions.
We propose and test the incorporation of two novel biases to enhance the search for effective counterfactual explanations.
arXiv Detail & Related papers (2024-04-19T11:47:17Z) - TextGenSHAP: Scalable Post-hoc Explanations in Text Generation with Long
Documents [34.52684986240312]
We introduce TextGenSHAP, an efficient post-hoc explanation method incorporating LM-specific techniques.
We demonstrate that this leads to significant increases in speed compared to conventional Shapley value computations.
In addition, we demonstrate how real-time Shapley values can be utilized in two important scenarios.
arXiv Detail & Related papers (2023-12-03T04:35:04Z) - Transcending Forgery Specificity with Latent Space Augmentation for Generalizable Deepfake Detection [57.646582245834324]
We propose a simple yet effective deepfake detector called LSDA.
It is based on a idea: representations with a wider variety of forgeries should be able to learn a more generalizable decision boundary.
We show that our proposed method is surprisingly effective and transcends state-of-the-art detectors across several widely used benchmarks.
arXiv Detail & Related papers (2023-11-19T09:41:10Z) - Consistent Multi-Granular Rationale Extraction for Explainable Multi-hop
Fact Verification [13.72453491358488]
This paper explores the viability of multi-granular rationale extraction with consistency and faithfulness for explainable multi-hop fact verification.
In particular, given a pretrained veracity prediction model, both the token-level explainer and sentence-level explainer are trained simultaneously to obtain multi-granular rationales.
Experimental results on three multi-hop fact verification datasets show that the proposed approach outperforms some state-of-the-art baselines.
arXiv Detail & Related papers (2023-05-16T12:31:53Z) - Improving Out-of-Distribution Generalization of Neural Rerankers with
Contextualized Late Interaction [52.63663547523033]
Late interaction, the simplest form of multi-vector, is also helpful to neural rerankers that only use the [] vector to compute the similarity score.
We show that the finding is consistent across different model sizes and first-stage retrievers of diverse natures.
arXiv Detail & Related papers (2023-02-13T18:42:17Z) - Explanation Selection Using Unlabeled Data for Chain-of-Thought
Prompting [80.9896041501715]
Explanations that have not been "tuned" for a task, such as off-the-shelf explanations written by nonexperts, may lead to mediocre performance.
This paper tackles the problem of how to optimize explanation-infused prompts in a blackbox fashion.
arXiv Detail & Related papers (2023-02-09T18:02:34Z) - Don't Explain Noise: Robust Counterfactuals for Randomized Ensembles [50.81061839052459]
We formalize the generation of robust counterfactual explanations as a probabilistic problem.
We show the link between the robustness of ensemble models and the robustness of base learners.
Our method achieves high robustness with only a small increase in the distance from counterfactual explanations to their initial observations.
arXiv Detail & Related papers (2022-05-27T17:28:54Z) - Latent-CF: A Simple Baseline for Reverse Counterfactual Explanations [0.0]
High quality explanations are first step in assessing fairness.
It is important to find a baseline for producing them.
We show that latent space counterfactual generation strikes balance between the speed of basic feature gradient descent methods and authenticity of counterfactuals generated by more complex feature space oriented techniques.
arXiv Detail & Related papers (2020-12-16T22:35:42Z) - SCOUT: Self-aware Discriminant Counterfactual Explanations [78.79534272979305]
The problem of counterfactual visual explanations is considered.
A new family of discriminant explanations is introduced.
The resulting counterfactual explanations are optimization free and thus much faster than previous methods.
arXiv Detail & Related papers (2020-04-16T17:05:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.