Let's Go to the Alien Zoo: Introducing an Experimental Framework to
Study Usability of Counterfactual Explanations for Machine Learning
- URL: http://arxiv.org/abs/2205.03398v1
- Date: Fri, 6 May 2022 17:57:05 GMT
- Title: Let's Go to the Alien Zoo: Introducing an Experimental Framework to
Study Usability of Counterfactual Explanations for Machine Learning
- Authors: Ulrike Kuhl and Andr\'e Artelt and Barbara Hammer
- Abstract summary: Counterfactual explanations (CFEs) have gained traction as a psychologically grounded approach to generate post-hoc explanations.
We introduce the Alien Zoo, an engaging, web-based and game-inspired experimental framework.
As a proof of concept, we demonstrate the practical efficacy and feasibility of this approach in a user study.
- Score: 6.883906273999368
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: To foster usefulness and accountability of machine learning (ML), it is
essential to explain a model's decisions in addition to evaluating its
performance. Accordingly, the field of explainable artificial intelligence
(XAI) has resurfaced as a topic of active research, offering approaches to
address the "how" and "why" of automated decision-making. Within this domain,
counterfactual explanations (CFEs) have gained considerable traction as a
psychologically grounded approach to generate post-hoc explanations. To do so,
CFEs highlight what changes to a model's input would have changed its
prediction in a particular way. However, despite the introduction of numerous
CFE approaches, their usability has yet to be thoroughly validated at the human
level. Thus, to advance the field of XAI, we introduce the Alien Zoo, an
engaging, web-based and game-inspired experimental framework. The Alien Zoo
provides the means to evaluate usability of CFEs for gaining new knowledge from
an automated system, targeting novice users in a domain-general context. As a
proof of concept, we demonstrate the practical efficacy and feasibility of this
approach in a user study. Our results suggest that users benefit from receiving
CFEs compared to no explanation, both in terms of objective performance in the
proposed iterative learning task, and subjective usability. With this work, we
aim to equip research groups and practitioners with the means to easily run
controlled and well-powered user studies to complement their otherwise often
more technology-oriented work. Thus, in the interest of reproducible research,
we provide the entire code, together with the underlying models and user data.
Related papers
- Introducing User Feedback-based Counterfactual Explanations (UFCE) [49.1574468325115]
Counterfactual explanations (CEs) have emerged as a viable solution for generating comprehensible explanations in XAI.
UFCE allows for the inclusion of user constraints to determine the smallest modifications in the subset of actionable features.
UFCE outperforms two well-known CE methods in terms of textitproximity, textitsparsity, and textitfeasibility.
arXiv Detail & Related papers (2024-02-26T20:09:44Z) - I-CEE: Tailoring Explanations of Image Classification Models to User
Expertise [13.293968260458962]
We present I-CEE, a framework that provides Image Classification Explanations tailored to User Expertise.
I-CEE models the informativeness of the example images to depend on user expertise, resulting in different examples for different users.
Experiments with simulated users show that I-CEE improves users' ability to accurately predict the model's decisions.
arXiv Detail & Related papers (2023-12-19T12:26:57Z) - Evaluating the Utility of Model Explanations for Model Development [54.23538543168767]
We evaluate whether explanations can improve human decision-making in practical scenarios of machine learning model development.
To our surprise, we did not find evidence of significant improvement on tasks when users were provided with any of the saliency maps.
These findings suggest caution regarding the usefulness and potential for misunderstanding in saliency-based explanations.
arXiv Detail & Related papers (2023-12-10T23:13:23Z) - RLIF: Interactive Imitation Learning as Reinforcement Learning [56.997263135104504]
We show how off-policy reinforcement learning can enable improved performance under assumptions that are similar but potentially even more practical than those of interactive imitation learning.
Our proposed method uses reinforcement learning with user intervention signals themselves as rewards.
This relaxes the assumption that intervening experts in interactive imitation learning should be near-optimal and enables the algorithm to learn behaviors that improve over the potential suboptimal human expert.
arXiv Detail & Related papers (2023-11-21T21:05:21Z) - Learning by Self-Explaining [23.420673675343266]
We introduce a novel workflow in the context of image classification, termed Learning by Self-Explaining (LSX)
LSX utilizes aspects of self-refining AI and human-guided explanatory machine learning.
Our results indicate improvements via Learning by Self-Explaining on several levels.
arXiv Detail & Related papers (2023-09-15T13:41:57Z) - Learning Action-Effect Dynamics for Hypothetical Vision-Language
Reasoning Task [50.72283841720014]
We propose a novel learning strategy that can improve reasoning about the effects of actions.
We demonstrate the effectiveness of our proposed approach and discuss its advantages over previous baselines in terms of performance, data efficiency, and generalization capability.
arXiv Detail & Related papers (2022-12-07T05:41:58Z) - Keep Your Friends Close and Your Counterfactuals Closer: Improved
Learning From Closest Rather Than Plausible Counterfactual Explanations in an
Abstract Setting [6.883906273999368]
Counterfactual explanations (CFEs) highlight what changes to a model's input would have changed its prediction in a particular way.
Recent innovations introduce the notion of computational plausibility for automatically generated CFEs.
We evaluate objective and subjective usability of computationally plausible CFEs in an iterative learning design targeting novice users.
arXiv Detail & Related papers (2022-05-11T14:07:57Z) - AcME -- Accelerated Model-agnostic Explanations: Fast Whitening of the
Machine-Learning Black Box [1.7534486934148554]
interpretability approaches should provide actionable insights without making the users wait.
We propose Accelerated Model-agnostic Explanations (AcME), an interpretability approach that quickly provides feature importance scores both at the global and the local level.
AcME computes feature ranking, but it also provides a what-if analysis tool to assess how changes in features values would affect model predictions.
arXiv Detail & Related papers (2021-12-23T15:18:13Z) - A general framework for scientifically inspired explanations in AI [76.48625630211943]
We instantiate the concept of structure of scientific explanation as the theoretical underpinning for a general framework in which explanations for AI systems can be implemented.
This framework aims to provide the tools to build a "mental-model" of any AI system so that the interaction with the user can provide information on demand and be closer to the nature of human-made explanations.
arXiv Detail & Related papers (2020-03-02T10:32:21Z) - Explainable Active Learning (XAL): An Empirical Study of How Local
Explanations Impact Annotator Experience [76.9910678786031]
We propose a novel paradigm of explainable active learning (XAL), by introducing techniques from the recently surging field of explainable AI (XAI) into an Active Learning setting.
Our study shows benefits of AI explanation as interfaces for machine teaching--supporting trust calibration and enabling rich forms of teaching feedback, and potential drawbacks--anchoring effect with the model judgment and cognitive workload.
arXiv Detail & Related papers (2020-01-24T22:52:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.