Causal Generative Explainers using Counterfactual Inference: A Case
Study on the Morpho-MNIST Dataset
- URL: http://arxiv.org/abs/2401.11394v1
- Date: Sun, 21 Jan 2024 04:07:48 GMT
- Title: Causal Generative Explainers using Counterfactual Inference: A Case
Study on the Morpho-MNIST Dataset
- Authors: Will Taylor-Melanson and Zahra Sadeghi and Stan Matwin
- Abstract summary: We present a generative counterfactual inference approach to study the influence of visual features as well as causal factors.
We employ visual explanation methods from OmnixAI open source toolkit to compare them with our proposed methods.
This finding suggests that our methods are well-suited for generating highly interpretable counterfactual explanations on causal datasets.
- Score: 5.458813674116228
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we propose leveraging causal generative learning as an
interpretable tool for explaining image classifiers. Specifically, we present a
generative counterfactual inference approach to study the influence of visual
features (i.e., pixels) as well as causal factors through generative learning.
To this end, we first uncover the most influential pixels on a classifier's
decision by varying the value of a causal attribute via counterfactual
inference and computing both Shapely and contrastive explanations for
counterfactual images with these different attribute values. We then establish
a Monte-Carlo mechanism using the generator of a causal generative model in
order to adapt Shapley explainers to produce feature importances for the
human-interpretable attributes of a causal dataset in the case where a
classifier has been trained exclusively on the images of the dataset. Finally,
we present optimization methods for creating counterfactual explanations of
classifiers by means of counterfactual inference, proposing straightforward
approaches for both differentiable and arbitrary classifiers. We exploit the
Morpho-MNIST causal dataset as a case study for exploring our proposed methods
for generating counterfacutl explantions. We employ visual explanation methods
from OmnixAI open source toolkit to compare them with our proposed methods. By
employing quantitative metrics to measure the interpretability of
counterfactual explanations, we find that our proposed methods of
counterfactual explanation offer more interpretable explanations compared to
those generated from OmnixAI. This finding suggests that our methods are
well-suited for generating highly interpretable counterfactual explanations on
causal datasets.
Related papers
- Rethinking Distance Metrics for Counterfactual Explainability [53.436414009687]
We investigate a framing for counterfactual generation methods that considers counterfactuals not as independent draws from a region around the reference, but as jointly sampled with the reference from the underlying data distribution.
We derive a distance metric, tailored for counterfactual similarity that can be applied to a broad range of settings.
arXiv Detail & Related papers (2024-10-18T15:06:50Z) - Selective Explanations [14.312717332216073]
A machine learning model is trained to predict feature attribution scores with only one inference.
Despite their efficiency, amortized explainers can produce inaccurate predictions and misleading explanations.
We propose selective explanations, a novel feature attribution method that detects when amortized explainers generate low-quality explanations.
arXiv Detail & Related papers (2024-05-29T23:08:31Z) - CNN-based explanation ensembling for dataset, representation and explanations evaluation [1.1060425537315088]
We explore the potential of ensembling explanations generated by deep classification models using convolutional model.
Through experimentation and analysis, we aim to investigate the implications of combining explanations to uncover a more coherent and reliable patterns of the model's behavior.
arXiv Detail & Related papers (2024-04-16T08:39:29Z) - Diffexplainer: Towards Cross-modal Global Explanations with Diffusion Models [51.21351775178525]
DiffExplainer is a novel framework that, leveraging language-vision models, enables multimodal global explainability.
It employs diffusion models conditioned on optimized text prompts, synthesizing images that maximize class outputs.
The analysis of generated visual descriptions allows for automatic identification of biases and spurious features.
arXiv Detail & Related papers (2024-04-03T10:11:22Z) - Viewing the process of generating counterfactuals as a source of knowledge: a new approach for explaining classifiers [0.0]
This article proposes to view this simulation process as a source of creating a certain amount of knowledge that can be stored to be used, later, in different ways.
arXiv Detail & Related papers (2023-09-08T12:06:48Z) - CLIMAX: An exploration of Classifier-Based Contrastive Explanations [5.381004207943597]
We propose a novel post-hoc model XAI technique that provides contrastive explanations justifying the classification of a black box.
Our method, which we refer to as CLIMAX, is based on local classifiers.
We show that we achieve better consistency as compared to baselines such as LIME, BayLIME, and SLIME.
arXiv Detail & Related papers (2023-07-02T22:52:58Z) - VCNet: A self-explaining model for realistic counterfactual generation [52.77024349608834]
Counterfactual explanation is a class of methods to make local explanations of machine learning decisions.
We present VCNet-Variational Counter Net, a model architecture that combines a predictor and a counterfactual generator.
We show that VCNet is able to both generate predictions, and to generate counterfactual explanations without having to solve another minimisation problem.
arXiv Detail & Related papers (2022-12-21T08:45:32Z) - Discriminative Attribution from Counterfactuals [64.94009515033984]
We present a method for neural network interpretability by combining feature attribution with counterfactual explanations.
We show that this method can be used to quantitatively evaluate the performance of feature attribution methods in an objective manner.
arXiv Detail & Related papers (2021-09-28T00:53:34Z) - Beyond Trivial Counterfactual Explanations with Diverse Valuable
Explanations [64.85696493596821]
In computer vision applications, generative counterfactual methods indicate how to perturb a model's input to change its prediction.
We propose a counterfactual method that learns a perturbation in a disentangled latent space that is constrained using a diversity-enforcing loss.
Our model improves the success rate of producing high-quality valuable explanations when compared to previous state-of-the-art methods.
arXiv Detail & Related papers (2021-03-18T12:57:34Z) - Generative Counterfactuals for Neural Networks via Attribute-Informed
Perturbation [51.29486247405601]
We design a framework to generate counterfactuals for raw data instances with the proposed Attribute-Informed Perturbation (AIP)
By utilizing generative models conditioned with different attributes, counterfactuals with desired labels can be obtained effectively and efficiently.
Experimental results on real-world texts and images demonstrate the effectiveness, sample quality as well as efficiency of our designed framework.
arXiv Detail & Related papers (2021-01-18T08:37:13Z) - Generative causal explanations of black-box classifiers [15.029443432414947]
We develop a method for generating causal post-hoc explanations of black-box classifiers based on a learned low-dimensional representation of the data.
We then demonstrate the practical utility of our method on image recognition tasks.
arXiv Detail & Related papers (2020-06-24T17:45:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.