A Few Good Counterfactuals: Generating Interpretable, Plausible and
Diverse Counterfactual Explanations
- URL: http://arxiv.org/abs/2101.09056v1
- Date: Fri, 22 Jan 2021 11:30:26 GMT
- Title: A Few Good Counterfactuals: Generating Interpretable, Plausible and
Diverse Counterfactual Explanations
- Authors: Barry Smyth and Mark T Keane
- Abstract summary: Good, native counterfactuals have been shown to rarely occur in most datasets.
Most popular methods generate synthetic counterfactuals using blind perturbations.
We describe a method that adapts native counterfactuals in the original dataset to generate sparse, diverse synthetic counterfactuals.
- Score: 14.283774141604997
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Counterfactual explanations provide a potentially significant solution to the
Explainable AI (XAI) problem, but good, native counterfactuals have been shown
to rarely occur in most datasets. Hence, the most popular methods generate
synthetic counterfactuals using blind perturbation. However, such methods have
several shortcomings: the resulting counterfactuals (i) may not be valid
data-points (they often use features that do not naturally occur), (ii) may
lack the sparsity of good counterfactuals (if they modify too many features),
and (iii) may lack diversity (if the generated counterfactuals are minimal
variants of one another). We describe a method designed to overcome these
problems, one that adapts native counterfactuals in the original dataset, to
generate sparse, diverse synthetic counterfactuals from naturally occurring
features. A series of experiments are reported that systematically explore
parametric variations of this novel method on common datasets to establish the
conditions for optimal performance.
Related papers
- Mining Invariance from Nonlinear Multi-Environment Data: Binary Classification [2.0528878959274883]
This paper focuses on binary classification to shed light on general nonlinear data generation mechanisms.
We identify a unique form of invariance that exists solely in a binary setting that allows us to train models invariant over environments.
We propose a prediction method and conduct experiments using real and synthetic datasets.
arXiv Detail & Related papers (2024-04-23T17:26:59Z) - Flexible and Robust Counterfactual Explanations with Minimal Satisfiable
Perturbations [56.941276017696076]
We propose a conceptually simple yet effective solution named Counterfactual Explanations with Minimal Satisfiable Perturbations (CEMSP)
CEMSP constrains changing values of abnormal features with the help of their semantically meaningful normal ranges.
Compared to existing methods, we conduct comprehensive experiments on both synthetic and real-world datasets to demonstrate that our method provides more robust explanations while preserving flexibility.
arXiv Detail & Related papers (2023-09-09T04:05:56Z) - CEnt: An Entropy-based Model-agnostic Explainability Framework to
Contrast Classifiers' Decisions [2.543865489517869]
We present a novel approach to locally contrast the prediction of any classifier.
Our Contrastive Entropy-based explanation method, CEnt, approximates a model locally by a decision tree to compute entropy information of different feature splits.
CEnt is the first non-gradient-based contrastive method generating diverse counterfactuals that do not necessarily exist in the training data while satisfying immutability (ex. race) and semi-immutability (ex. age can only change in an increasing direction)
arXiv Detail & Related papers (2023-01-19T08:23:34Z) - Posterior Collapse and Latent Variable Non-identifiability [54.842098835445]
We propose a class of latent-identifiable variational autoencoders, deep generative models which enforce identifiability without sacrificing flexibility.
Across synthetic and real datasets, latent-identifiable variational autoencoders outperform existing methods in mitigating posterior collapse and providing meaningful representations of the data.
arXiv Detail & Related papers (2023-01-02T06:16:56Z) - MACE: An Efficient Model-Agnostic Framework for Counterfactual
Explanation [132.77005365032468]
We propose a novel framework of Model-Agnostic Counterfactual Explanation (MACE)
In our MACE approach, we propose a novel RL-based method for finding good counterfactual examples and a gradient-less descent method for improving proximity.
Experiments on public datasets validate the effectiveness with better validity, sparsity and proximity.
arXiv Detail & Related papers (2022-05-31T04:57:06Z) - Equivariance Allows Handling Multiple Nuisance Variables When Analyzing
Pooled Neuroimaging Datasets [53.34152466646884]
In this paper, we show how bringing recent results on equivariant representation learning instantiated on structured spaces together with simple use of classical results on causal inference provides an effective practical solution.
We demonstrate how our model allows dealing with more than one nuisance variable under some assumptions and can enable analysis of pooled scientific datasets in scenarios that would otherwise entail removing a large portion of the samples.
arXiv Detail & Related papers (2022-03-29T04:54:06Z) - Agree to Disagree: Diversity through Disagreement for Better
Transferability [54.308327969778155]
We propose D-BAT (Diversity-By-disAgreement Training), which enforces agreement among the models on the training data.
We show how D-BAT naturally emerges from the notion of generalized discrepancy.
arXiv Detail & Related papers (2022-02-09T12:03:02Z) - LARD: Large-scale Artificial Disfluency Generation [0.0]
We propose LARD, a method for generating complex and realistic artificial disfluencies with little effort.
The proposed method can handle three of the most common types of disfluencies: repetitions, replacements and restarts.
We release a new large-scale dataset with disfluencies that can be used on four different tasks.
arXiv Detail & Related papers (2022-01-13T16:02:36Z) - Generalization of Neural Combinatorial Solvers Through the Lens of
Adversarial Robustness [68.97830259849086]
Most datasets only capture a simpler subproblem and likely suffer from spurious features.
We study adversarial robustness - a local generalization property - to reveal hard, model-specific instances and spurious features.
Unlike in other applications, where perturbation models are designed around subjective notions of imperceptibility, our perturbation models are efficient and sound.
Surprisingly, with such perturbations, a sufficiently expressive neural solver does not suffer from the limitations of the accuracy-robustness trade-off common in supervised learning.
arXiv Detail & Related papers (2021-10-21T07:28:11Z) - Causality-based Counterfactual Explanation for Classification Models [11.108866104714627]
We propose a prototype-based counterfactual explanation framework (ProCE)
ProCE is capable of preserving the causal relationship underlying the features of the counterfactual data.
In addition, we design a novel gradient-free optimization based on the multi-objective genetic algorithm that generates the counterfactual explanations.
arXiv Detail & Related papers (2021-05-03T09:25:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.