A Symbolic Approach for Counterfactual Explanations
- URL: http://arxiv.org/abs/2206.09638v1
- Date: Mon, 20 Jun 2022 08:38:54 GMT
- Title: A Symbolic Approach for Counterfactual Explanations
- Authors: Ryma Boumazouza (UA, CNRS, CRIL), Fahima Cheikh-Alili (UA, CNRS,
CRIL), Bertrand Mazure (UA, CNRS, CRIL), Karim Tabia (UA, CNRS, CRIL)
- Abstract summary: We propose a novel symbolic approach to provide counterfactual explanations for a classifier predictions.
Our approach is symbolic in the sense that it is based on encoding the decision function of a classifier in an equivalent CNF formula.
- Score: 18.771531343438227
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper titled A Symbolic Approach for Counterfactual Explanations we
propose a novel symbolic approach to provide counterfactual explanations for a
classifier predictions. Contrary to most explanation approaches where the goal
is to understand which and to what extent parts of the data helped to give a
prediction, counterfactual explanations indicate which features must be changed
in the data in order to change this classifier prediction. Our approach is
symbolic in the sense that it is based on encoding the decision function of a
classifier in an equivalent CNF formula. In this approach, counterfactual
explanations are seen as the Minimal Correction Subsets (MCS), a well-known
concept in knowledge base reparation. Hence, this approach takes advantage of
the strengths of already existing and proven solutions for the generation of
MCS. Our preliminary experimental studies on Bayesian classifiers show the
potential of this approach on several datasets.
Related papers
- Rethinking Distance Metrics for Counterfactual Explainability [53.436414009687]
We investigate a framing for counterfactual generation methods that considers counterfactuals not as independent draws from a region around the reference, but as jointly sampled with the reference from the underlying data distribution.
We derive a distance metric, tailored for counterfactual similarity that can be applied to a broad range of settings.
arXiv Detail & Related papers (2024-10-18T15:06:50Z) - An Axiomatic Approach to Model-Agnostic Concept Explanations [67.84000759813435]
We propose an approach to concept explanations that satisfy three natural axioms: linearity, recursivity, and similarity.
We then establish connections with previous concept explanation methods, offering insight into their varying semantic meanings.
arXiv Detail & Related papers (2024-01-12T20:53:35Z) - Viewing the process of generating counterfactuals as a source of knowledge: a new approach for explaining classifiers [0.0]
This article proposes to view this simulation process as a source of creating a certain amount of knowledge that can be stored to be used, later, in different ways.
arXiv Detail & Related papers (2023-09-08T12:06:48Z) - ASTERYX : A model-Agnostic SaT-basEd appRoach for sYmbolic and
score-based eXplanations [26.500149465292246]
This paper proposes a generic approach named ASTERYX allowing to generate both symbolic explanations and score-based ones.
Our experimental results show the feasibility of the proposed approach and its effectiveness in providing symbolic and score-based explanations.
arXiv Detail & Related papers (2022-06-23T08:37:32Z) - A Model-Agnostic SAT-based Approach for Symbolic Explanation Enumeration [26.500149465292246]
We generate explanations to locally explain a single prediction by analyzing the relationship between the features and the output.
Our approach uses a propositional encoding of the predictive model and a SAT-based setting to generate two types of symbolic explanations Sufficient Reasons and Counterfactuals.
arXiv Detail & Related papers (2022-06-23T08:35:47Z) - Logical Satisfiability of Counterfactuals for Faithful Explanations in
NLI [60.142926537264714]
We introduce the methodology of Faithfulness-through-Counterfactuals.
It generates a counterfactual hypothesis based on the logical predicates expressed in the explanation.
It then evaluates if the model's prediction on the counterfactual is consistent with that expressed logic.
arXiv Detail & Related papers (2022-05-25T03:40:59Z) - Provable concept learning for interpretable predictions using
variational inference [7.0349768355860895]
In safety critical applications, practitioners are reluctant to trust neural networks when no interpretable explanations are available.
We propose a probabilistic modeling framework to derive (C)oncept (L)earning and (P)rediction (CLAP)
We prove that our method is able to identify them while attaining optimal classification accuracy.
arXiv Detail & Related papers (2022-04-01T14:51:38Z) - Resolving label uncertainty with implicit posterior models [71.62113762278963]
We propose a method for jointly inferring labels across a collection of data samples.
By implicitly assuming the existence of a generative model for which a differentiable predictor is the posterior, we derive a training objective that allows learning under weak beliefs.
arXiv Detail & Related papers (2022-02-28T18:09:44Z) - Discriminative Attribution from Counterfactuals [64.94009515033984]
We present a method for neural network interpretability by combining feature attribution with counterfactual explanations.
We show that this method can be used to quantitatively evaluate the performance of feature attribution methods in an objective manner.
arXiv Detail & Related papers (2021-09-28T00:53:34Z) - EDDA: Explanation-driven Data Augmentation to Improve Model and
Explanation Alignment [12.729179495550557]
We seek a methodology that can improve alignment between model predictions and explanation method.
We achieve this through a novel explanation-driven data augmentation (EDDA) method.
This is based on the simple motivating principle that occluding salient regions for the model prediction should decrease the model confidence in the prediction.
arXiv Detail & Related papers (2021-05-29T00:42:42Z) - Evaluations and Methods for Explanation through Robustness Analysis [117.7235152610957]
We establish a novel set of evaluation criteria for such feature based explanations by analysis.
We obtain new explanations that are loosely necessary and sufficient for a prediction.
We extend the explanation to extract the set of features that would move the current prediction to a target class.
arXiv Detail & Related papers (2020-05-31T05:52:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.