Keep Your Friends Close and Your Counterfactuals Closer: Improved
Learning From Closest Rather Than Plausible Counterfactual Explanations in an
Abstract Setting
- URL: http://arxiv.org/abs/2205.05515v1
- Date: Wed, 11 May 2022 14:07:57 GMT
- Title: Keep Your Friends Close and Your Counterfactuals Closer: Improved
Learning From Closest Rather Than Plausible Counterfactual Explanations in an
Abstract Setting
- Authors: Ulrike Kuhl and Andr\'e Artelt and Barbara Hammer
- Abstract summary: Counterfactual explanations (CFEs) highlight what changes to a model's input would have changed its prediction in a particular way.
Recent innovations introduce the notion of computational plausibility for automatically generated CFEs.
We evaluate objective and subjective usability of computationally plausible CFEs in an iterative learning design targeting novice users.
- Score: 6.883906273999368
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Counterfactual explanations (CFEs) highlight what changes to a model's input
would have changed its prediction in a particular way. CFEs have gained
considerable traction as a psychologically grounded solution for explainable
artificial intelligence (XAI). Recent innovations introduce the notion of
computational plausibility for automatically generated CFEs, enhancing their
robustness by exclusively creating plausible explanations. However, practical
benefits of such a constraint on user experience and behavior is yet unclear.
In this study, we evaluate objective and subjective usability of
computationally plausible CFEs in an iterative learning design targeting novice
users. We rely on a novel, game-like experimental design, revolving around an
abstract scenario. Our results show that novice users actually benefit less
from receiving computationally plausible rather than closest CFEs that produce
minimal changes leading to the desired outcome. Responses in a post-game survey
reveal no differences in terms of subjective user experience between both
groups. Following the view of psychological plausibility as comparative
similarity, this may be explained by the fact that users in the closest
condition experience their CFEs as more psychologically plausible than the
computationally plausible counterpart. In sum, our work highlights a
little-considered divergence of definitions of computational plausibility and
psychological plausibility, critically confirming the need to incorporate human
behavior, preferences and mental models already at the design stages of XAI
approaches. In the interest of reproducible research, all source code, acquired
user data, and evaluation scripts of the current study are available:
https://github.com/ukuhl/PlausibleAlienZoo
Related papers
- Introducing User Feedback-based Counterfactual Explanations (UFCE) [49.1574468325115]
Counterfactual explanations (CEs) have emerged as a viable solution for generating comprehensible explanations in XAI.
UFCE allows for the inclusion of user constraints to determine the smallest modifications in the subset of actionable features.
UFCE outperforms two well-known CE methods in terms of textitproximity, textitsparsity, and textitfeasibility.
arXiv Detail & Related papers (2024-02-26T20:09:44Z) - LaPLACE: Probabilistic Local Model-Agnostic Causal Explanations [1.0370398945228227]
We introduce LaPLACE-explainer, designed to provide probabilistic cause-and-effect explanations for machine learning models.
The LaPLACE-Explainer component leverages the concept of a Markov blanket to establish statistical boundaries between relevant and non-relevant features.
Our approach offers causal explanations and outperforms LIME and SHAP in terms of local accuracy and consistency of explained features.
arXiv Detail & Related papers (2023-10-01T04:09:59Z) - Flexible and Robust Counterfactual Explanations with Minimal Satisfiable
Perturbations [56.941276017696076]
We propose a conceptually simple yet effective solution named Counterfactual Explanations with Minimal Satisfiable Perturbations (CEMSP)
CEMSP constrains changing values of abnormal features with the help of their semantically meaningful normal ranges.
Compared to existing methods, we conduct comprehensive experiments on both synthetic and real-world datasets to demonstrate that our method provides more robust explanations while preserving flexibility.
arXiv Detail & Related papers (2023-09-09T04:05:56Z) - For Better or Worse: The Impact of Counterfactual Explanations'
Directionality on User Behavior in xAI [6.883906273999368]
Counterfactual explanations (CFEs) are a popular approach in explainable artificial intelligence (xAI)
CFEs describe a scenario that is better than the factual state (upward CFE), or a scenario that is worse than the factual state (downward CFE)
This study compares the impact of CFE directionality on behavior and experience of participants tasked to extract new knowledge from an automated system.
arXiv Detail & Related papers (2023-06-13T09:16:38Z) - ASPEST: Bridging the Gap Between Active Learning and Selective
Prediction [56.001808843574395]
Selective prediction aims to learn a reliable model that abstains from making predictions when uncertain.
Active learning aims to lower the overall labeling effort, and hence human dependence, by querying the most informative examples.
In this work, we introduce a new learning paradigm, active selective prediction, which aims to query more informative samples from the shifted target domain.
arXiv Detail & Related papers (2023-04-07T23:51:07Z) - VCNet: A self-explaining model for realistic counterfactual generation [52.77024349608834]
Counterfactual explanation is a class of methods to make local explanations of machine learning decisions.
We present VCNet-Variational Counter Net, a model architecture that combines a predictor and a counterfactual generator.
We show that VCNet is able to both generate predictions, and to generate counterfactual explanations without having to solve another minimisation problem.
arXiv Detail & Related papers (2022-12-21T08:45:32Z) - Scaling Laws Beyond Backpropagation [64.0476282000118]
We study the ability of Direct Feedback Alignment to train causal decoder-only Transformers efficiently.
We find that DFA fails to offer more efficient scaling than backpropagation.
arXiv Detail & Related papers (2022-10-26T10:09:14Z) - Meta-Wrapper: Differentiable Wrapping Operator for User Interest
Selection in CTR Prediction [97.99938802797377]
Click-through rate (CTR) prediction, whose goal is to predict the probability of the user to click on an item, has become increasingly significant in recommender systems.
Recent deep learning models with the ability to automatically extract the user interest from his/her behaviors have achieved great success.
We propose a novel approach under the framework of the wrapper method, which is named Meta-Wrapper.
arXiv Detail & Related papers (2022-06-28T03:28:15Z) - Let's Go to the Alien Zoo: Introducing an Experimental Framework to
Study Usability of Counterfactual Explanations for Machine Learning [6.883906273999368]
Counterfactual explanations (CFEs) have gained traction as a psychologically grounded approach to generate post-hoc explanations.
We introduce the Alien Zoo, an engaging, web-based and game-inspired experimental framework.
As a proof of concept, we demonstrate the practical efficacy and feasibility of this approach in a user study.
arXiv Detail & Related papers (2022-05-06T17:57:05Z) - DISSECT: Disentangled Simultaneous Explanations via Concept Traversals [33.65478845353047]
DISSECT is a novel approach to explaining deep learning model inferences.
By training a generative model from a classifier's signal, DISSECT offers a way to discover a classifier's inherent "notion" of distinct concepts.
We show that DISSECT produces CTs that disentangle several concepts and are coupled to its reasoning due to joint training.
arXiv Detail & Related papers (2021-05-31T17:11:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.