Latent-CF: A Simple Baseline for Reverse Counterfactual Explanations
- URL: http://arxiv.org/abs/2012.09301v1
- Date: Wed, 16 Dec 2020 22:35:42 GMT
- Title: Latent-CF: A Simple Baseline for Reverse Counterfactual Explanations
- Authors: Rachana Balasubramanian, Samuel Sharpe, Brian Barr, Jason Wittenbach,
and C. Bayan Bruss
- Abstract summary: High quality explanations are first step in assessing fairness.
It is important to find a baseline for producing them.
We show that latent space counterfactual generation strikes balance between the speed of basic feature gradient descent methods and authenticity of counterfactuals generated by more complex feature space oriented techniques.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In the environment of fair lending laws and the General Data Protection
Regulation (GDPR), the ability to explain a model's prediction is of paramount
importance. High quality explanations are the first step in assessing fairness.
Counterfactuals are valuable tools for explainability. They provide actionable,
comprehensible explanations for the individual who is subject to decisions made
from the prediction. It is important to find a baseline for producing them. We
propose a simple method for generating counterfactuals by using gradient
descent to search in the latent space of an autoencoder and benchmark our
method against approaches that search for counterfactuals in feature space.
Additionally, we implement metrics to concretely evaluate the quality of the
counterfactuals. We show that latent space counterfactual generation strikes a
balance between the speed of basic feature gradient descent methods and the
sparseness and authenticity of counterfactuals generated by more complex
feature space oriented techniques.
Related papers
- Fast Calibrated Explanations: Efficient and Uncertainty-Aware Explanations for Machine Learning Models [41.82622187379551]
This paper introduces Fast Calibrated Explanations, a method for generating rapid, uncertainty-aware explanations for machine learning models.
By incorporating perturbation techniques from ConformaSight into the core elements of Calibrated Explanations, we achieve significant speedups.
While the new method sacrifices a small degree of detail, it excels in computational efficiency, making it ideal for high-stakes, real-time applications.
arXiv Detail & Related papers (2024-10-28T15:29:35Z) - Backdoor-based Explainable AI Benchmark for High Fidelity Evaluation of Attribution Methods [49.62131719441252]
Attribution methods compute importance scores for input features to explain the output predictions of deep models.
In this work, we first identify a set of fidelity criteria that reliable benchmarks for attribution methods are expected to fulfill.
We then introduce a Backdoor-based eXplainable AI benchmark (BackX) that adheres to the desired fidelity criteria.
arXiv Detail & Related papers (2024-05-02T13:48:37Z) - Faithful Explanations of Black-box NLP Models Using LLM-generated
Counterfactuals [67.64770842323966]
Causal explanations of predictions of NLP systems are essential to ensure safety and establish trust.
Existing methods often fall short of explaining model predictions effectively or efficiently.
We propose two approaches for counterfactual (CF) approximation.
arXiv Detail & Related papers (2023-10-01T07:31:04Z) - LaPLACE: Probabilistic Local Model-Agnostic Causal Explanations [1.0370398945228227]
We introduce LaPLACE-explainer, designed to provide probabilistic cause-and-effect explanations for machine learning models.
The LaPLACE-Explainer component leverages the concept of a Markov blanket to establish statistical boundaries between relevant and non-relevant features.
Our approach offers causal explanations and outperforms LIME and SHAP in terms of local accuracy and consistency of explained features.
arXiv Detail & Related papers (2023-10-01T04:09:59Z) - Simple Steps to Success: A Method for Step-Based Counterfactual Explanations [9.269923473051138]
We propose a data-driven and model-agnostic framework to compute counterfactual explanations.
We introduce StEP, a computationally efficient method that offers incremental steps along the data manifold that directs users towards their desired outcome.
arXiv Detail & Related papers (2023-06-27T15:35:22Z) - MACE: An Efficient Model-Agnostic Framework for Counterfactual
Explanation [132.77005365032468]
We propose a novel framework of Model-Agnostic Counterfactual Explanation (MACE)
In our MACE approach, we propose a novel RL-based method for finding good counterfactual examples and a gradient-less descent method for improving proximity.
Experiments on public datasets validate the effectiveness with better validity, sparsity and proximity.
arXiv Detail & Related papers (2022-05-31T04:57:06Z) - Don't Explain Noise: Robust Counterfactuals for Randomized Ensembles [50.81061839052459]
We formalize the generation of robust counterfactual explanations as a probabilistic problem.
We show the link between the robustness of ensemble models and the robustness of base learners.
Our method achieves high robustness with only a small increase in the distance from counterfactual explanations to their initial observations.
arXiv Detail & Related papers (2022-05-27T17:28:54Z) - Logical Satisfiability of Counterfactuals for Faithful Explanations in
NLI [60.142926537264714]
We introduce the methodology of Faithfulness-through-Counterfactuals.
It generates a counterfactual hypothesis based on the logical predicates expressed in the explanation.
It then evaluates if the model's prediction on the counterfactual is consistent with that expressed logic.
arXiv Detail & Related papers (2022-05-25T03:40:59Z) - Optimal Counterfactual Explanations in Tree Ensembles [3.8073142980733]
We advocate for a model-based search aiming at "optimal" explanations and propose efficient mixed-integer programming approaches.
We show that isolation forests can be modeled within our framework to focus the search on plausible explanations with a low outlier score.
arXiv Detail & Related papers (2021-06-11T22:44:27Z) - Memory networks for consumer protection:unfairness exposed [12.884439911728112]
Data-driven AI methods can leverage consumer protection by supporting the automated analysis of legal documents.
We consider several configurations of memory-augmented neural networks where rationales are given a special role in the modeling of context knowledge.
Our results show that rationales not only contribute to improve the classification accuracy, but are also able to offer meaningful, natural language explanations.
arXiv Detail & Related papers (2020-07-24T14:25:54Z) - Evaluations and Methods for Explanation through Robustness Analysis [117.7235152610957]
We establish a novel set of evaluation criteria for such feature based explanations by analysis.
We obtain new explanations that are loosely necessary and sufficient for a prediction.
We extend the explanation to extract the set of features that would move the current prediction to a target class.
arXiv Detail & Related papers (2020-05-31T05:52:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.