Abductive explanations of classifiers under constraints: Complexity and properties
- URL: http://arxiv.org/abs/2409.12154v1
- Date: Wed, 18 Sep 2024 17:15:39 GMT
- Title: Abductive explanations of classifiers under constraints: Complexity and properties
- Authors: Martin Cooper, Leila Amgoud,
- Abstract summary: We propose three new types of explanations that take into account constraints.
They can be generated from the whole feature space or from a dataset.
We show that coverage is powerful enough to discard redundant and superfluous AXp's.
- Score: 6.629765271909503
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Abductive explanations (AXp's) are widely used for understanding decisions of classifiers. Existing definitions are suitable when features are independent. However, we show that ignoring constraints when they exist between features may lead to an explosion in the number of redundant or superfluous AXp's. We propose three new types of explanations that take into account constraints and that can be generated from the whole feature space or from a sample (such as a dataset). They are based on a key notion of coverage of an explanation, the set of instances it explains. We show that coverage is powerful enough to discard redundant and superfluous AXp's. For each type, we analyse the complexity of finding an explanation and investigate its formal properties. The final result is a catalogue of different forms of AXp's with different complexities and different formal guarantees.
Related papers
- Axiomatic Characterisations of Sample-based Explainers [8.397730500554047]
We scrutinize explainers that generate feature-based explanations from samples or datasets.
We identify the entire family of explainers that satisfy two key properties which are compatible with all the others.
We introduce the first (broad family of) explainers that guarantee the existence of explanations and irrefutable global consistency.
arXiv Detail & Related papers (2024-08-09T07:10:07Z) - Ensemble of Counterfactual Explainers [17.88531216690148]
We propose an ensemble of counterfactual explainers that boosts weak explainers, which provide only a subset of such properties.
The ensemble runs weak explainers on a sample of instances and of features, and it combines their results by exploiting a diversity-driven selection function.
arXiv Detail & Related papers (2023-08-29T10:21:50Z) - Explanation Selection Using Unlabeled Data for Chain-of-Thought
Prompting [80.9896041501715]
Explanations that have not been "tuned" for a task, such as off-the-shelf explanations written by nonexperts, may lead to mediocre performance.
This paper tackles the problem of how to optimize explanation-infused prompts in a blackbox fashion.
arXiv Detail & Related papers (2023-02-09T18:02:34Z) - On Computing Probabilistic Abductive Explanations [30.325691263226968]
The most widely studied explainable AI (XAI) approaches are unsound.
PI-explanations also exhibit important drawbacks, the most visible of which is arguably their size.
This paper investigates practical approaches for computing relevant sets for a number of widely used classifiers.
arXiv Detail & Related papers (2022-12-12T15:47:10Z) - ELUDE: Generating interpretable explanations via a decomposition into
labelled and unlabelled features [23.384134043048807]
We develop an explanation framework that decomposes a model's prediction into two parts.
By identifying the latter, we are able to analyze the "unexplained" portion of the model.
We show that the set of unlabelled features can generalize to multiple models trained with the same feature space.
arXiv Detail & Related papers (2022-06-15T17:36:55Z) - Don't Explain Noise: Robust Counterfactuals for Randomized Ensembles [50.81061839052459]
We formalize the generation of robust counterfactual explanations as a probabilistic problem.
We show the link between the robustness of ensemble models and the robustness of base learners.
Our method achieves high robustness with only a small increase in the distance from counterfactual explanations to their initial observations.
arXiv Detail & Related papers (2022-05-27T17:28:54Z) - Contrastive Explanations for Model Interpretability [77.92370750072831]
We propose a methodology to produce contrastive explanations for classification models.
Our method is based on projecting model representation to a latent space.
Our findings shed light on the ability of label-contrastive explanations to provide a more accurate and finer-grained interpretability of a model's decision.
arXiv Detail & Related papers (2021-03-02T00:36:45Z) - Dependency Decomposition and a Reject Option for Explainable Models [4.94950858749529]
Recent deep learning models perform extremely well in various inference tasks.
Recent advances offer methods to visualize features, describe attribution of the input.
We present the first analysis of dependencies regarding the probability distribution over the desired image classification outputs.
arXiv Detail & Related papers (2020-12-11T17:39:33Z) - The Struggles of Feature-Based Explanations: Shapley Values vs. Minimal
Sufficient Subsets [61.66584140190247]
We show that feature-based explanations pose problems even for explaining trivial models.
We show that two popular classes of explainers, Shapley explainers and minimal sufficient subsets explainers, target fundamentally different types of ground-truth explanations.
arXiv Detail & Related papers (2020-09-23T09:45:23Z) - Robust Question Answering Through Sub-part Alignment [53.94003466761305]
We model question answering as an alignment problem.
We train our model on SQuAD v1.1 and test it on several adversarial and out-of-domain datasets.
arXiv Detail & Related papers (2020-04-30T09:10:57Z) - A Formal Approach to Explainability [100.12889473240237]
We study the links between explanation-generating functions and intermediate representations of learned models.
We study the intersection and union of explanations as a way to construct new explanations.
arXiv Detail & Related papers (2020-01-15T10:06:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.