Explaining $\mathcal{ELH}$ Concept Descriptions through Counterfactual
Reasoning
- URL: http://arxiv.org/abs/2301.05109v2
- Date: Wed, 4 Oct 2023 14:33:47 GMT
- Title: Explaining $\mathcal{ELH}$ Concept Descriptions through Counterfactual
Reasoning
- Authors: Leonie Nora Sieger, Stefan Heindorf, Yasir Mahmood, Lukas Bl\"ubaum,
Axel-Cyrille Ngonga Ngomo
- Abstract summary: An intrinsically transparent way to do classification is by using concepts in description logics.
One solution is to employ counterfactuals to answer the question, How must feature values be changed to obtain a different classification?''
- Score: 3.5323691899538128
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Knowledge bases are widely used for information management, enabling
high-impact applications such as web search, question answering, and natural
language processing. They also serve as the backbone for automatic decision
systems, e.g., for medical diagnostics and credit scoring. As stakeholders
affected by these decisions would like to understand their situation and verify
how fair the decisions are, a number of explanation approaches have been
proposed. An intrinsically transparent way to do classification is by using
concepts in description logics. However, these concepts can become long and
difficult to fathom for non-experts, even when verbalized. One solution is to
employ counterfactuals to answer the question, ``How must feature values be
changed to obtain a different classification?'' By focusing on the minimal
feature changes, the explanations are short, human-friendly, and provide a
clear path of action regarding the change in prediction. While previous work
investigated counterfactuals for tabular data, in this paper, we transfer the
notion of counterfactuals to knowledge bases and the description logic
$\mathcal{ELH}$. Our approach starts by generating counterfactual candidates
from concepts, followed by selecting the candidates requiring the fewest
feature changes as counterfactuals. When multiple counterfactuals exist, we
rank them based on the likeliness of their feature combinations. We evaluate
our method by conducting a user survey to determine which counterfactual
candidates participants prefer for explanation.
Related papers
- Relevant Irrelevance: Generating Alterfactual Explanations for Image Classifiers [11.200613814162185]
In this paper, we demonstrate the feasibility of alterfactual explanations for black box image classifiers.
We show for the first time that it is possible to apply this idea to black box models based on neural networks.
arXiv Detail & Related papers (2024-05-08T11:03:22Z) - Disagreement amongst counterfactual explanations: How transparency can
be deceptive [0.0]
Counterfactual explanations are increasingly used as Explainable Artificial Intelligence technique.
Not every algorithm creates uniform explanations for the same instance.
Ethical issues arise when malicious agents use this diversity to fairwash an unfair machine learning model.
arXiv Detail & Related papers (2023-04-25T09:15:37Z) - Explanation Selection Using Unlabeled Data for Chain-of-Thought
Prompting [80.9896041501715]
Explanations that have not been "tuned" for a task, such as off-the-shelf explanations written by nonexperts, may lead to mediocre performance.
This paper tackles the problem of how to optimize explanation-infused prompts in a blackbox fashion.
arXiv Detail & Related papers (2023-02-09T18:02:34Z) - Motivating explanations in Bayesian networks using MAP-independence [0.0]
In Bayesian networks a diagnosis or classification is typically formalized as the computation of the most probable joint value assignment to the hypothesis variables.
In this paper we introduce a new concept, MAP- independence, which tries to capture this notion of relevance.
arXiv Detail & Related papers (2022-08-05T12:26:54Z) - Features of Explainability: How users understand counterfactual and
causal explanations for categorical and continuous features in XAI [10.151828072611428]
Counterfactual explanations are increasingly used to address interpretability, recourse, and bias in AI decisions.
We tested the effects of counterfactual and causal explanations on the objective accuracy of users predictions.
We also found that users understand explanations referring to categorical features more readily than those referring to continuous features.
arXiv Detail & Related papers (2022-04-21T15:01:09Z) - Search Methods for Sufficient, Socially-Aligned Feature Importance
Explanations with In-Distribution Counterfactuals [72.00815192668193]
Feature importance (FI) estimates are a popular form of explanation, and they are commonly created and evaluated by computing the change in model confidence caused by removing certain input features at test time.
We study several under-explored dimensions of FI-based explanations, providing conceptual and empirical improvements for this form of explanation.
arXiv Detail & Related papers (2021-06-01T20:36:48Z) - Contrastive Explanations for Model Interpretability [77.92370750072831]
We propose a methodology to produce contrastive explanations for classification models.
Our method is based on projecting model representation to a latent space.
Our findings shed light on the ability of label-contrastive explanations to provide a more accurate and finer-grained interpretability of a model's decision.
arXiv Detail & Related papers (2021-03-02T00:36:45Z) - This is not the Texture you are looking for! Introducing Novel
Counterfactual Explanations for Non-Experts using Generative Adversarial
Learning [59.17685450892182]
counterfactual explanation systems try to enable a counterfactual reasoning by modifying the input image.
We present a novel approach to generate such counterfactual image explanations based on adversarial image-to-image translation techniques.
Our results show that our approach leads to significantly better results regarding mental models, explanation satisfaction, trust, emotions, and self-efficacy than two state-of-the art systems.
arXiv Detail & Related papers (2020-12-22T10:08:05Z) - SCOUT: Self-aware Discriminant Counterfactual Explanations [78.79534272979305]
The problem of counterfactual visual explanations is considered.
A new family of discriminant explanations is introduced.
The resulting counterfactual explanations are optimization free and thus much faster than previous methods.
arXiv Detail & Related papers (2020-04-16T17:05:49Z) - An Information-Theoretic Approach to Personalized Explainable Machine
Learning [92.53970625312665]
We propose a simple probabilistic model for the predictions and user knowledge.
We quantify the effect of an explanation by the conditional mutual information between the explanation and prediction.
arXiv Detail & Related papers (2020-03-01T13:06:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.