DisCERN:Discovering Counterfactual Explanations using Relevance Features
from Neighbourhoods
- URL: http://arxiv.org/abs/2109.05800v1
- Date: Mon, 13 Sep 2021 09:25:25 GMT
- Title: DisCERN:Discovering Counterfactual Explanations using Relevance Features
from Neighbourhoods
- Authors: Nirmalie Wiratunga, Anjana Wijekoon, Ikechukwu Nkisi-Orji, Kyle
Martin, Chamath Palihawadana, David Corsar
- Abstract summary: DisCERN is an effective strategy to minimise actionable changes necessary to create good counterfactual explanations.
We show how widely adopted feature relevance-based explainers can inform DisCERN to identify the minimum subset of "actionable features"
Our results demonstrate that DisCERN is an effective strategy to minimise actionable changes necessary to create good counterfactual explanations.
- Score: 1.9706200133168679
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Counterfactual explanations focus on "actionable knowledge" to help end-users
understand how a machine learning outcome could be changed to a more desirable
outcome. For this purpose a counterfactual explainer needs to discover input
dependencies that relate to outcome changes. Identifying the minimum subset of
feature changes needed to action an output change in the decision is an
interesting challenge for counterfactual explainers. The DisCERN algorithm
introduced in this paper is a case-based counter-factual explainer. Here
counterfactuals are formed by replacing feature values from a nearest unlike
neighbour (NUN) until an actionable change is observed. We show how widely
adopted feature relevance-based explainers (i.e. LIME, SHAP), can inform
DisCERN to identify the minimum subset of "actionable features". We demonstrate
our DisCERN algorithm on five datasets in a comparative study with the widely
used optimisation-based counterfactual approach DiCE. Our results demonstrate
that DisCERN is an effective strategy to minimise actionable changes necessary
to create good counterfactual explanations.
Related papers
- Using LLMs for Explaining Sets of Counterfactual Examples to Final Users [0.0]
In automated decision-making scenarios, causal inference methods can analyze the underlying data-generation process.
Counterfactual examples explore hypothetical scenarios where a minimal number of factors are altered.
We propose a novel multi-step pipeline that uses counterfactuals to generate natural language explanations of actions that will lead to a change in outcome.
arXiv Detail & Related papers (2024-08-27T15:13:06Z) - Introducing User Feedback-based Counterfactual Explanations (UFCE) [49.1574468325115]
Counterfactual explanations (CEs) have emerged as a viable solution for generating comprehensible explanations in XAI.
UFCE allows for the inclusion of user constraints to determine the smallest modifications in the subset of actionable features.
UFCE outperforms two well-known CE methods in terms of textitproximity, textitsparsity, and textitfeasibility.
arXiv Detail & Related papers (2024-02-26T20:09:44Z) - Generating robust counterfactual explanations [60.32214822437734]
The quality of a counterfactual depends on several criteria: realism, actionability, validity, robustness, etc.
In this paper, we are interested in the notion of robustness of a counterfactual. More precisely, we focus on robustness to counterfactual input changes.
We propose a new framework, CROCO, that generates robust counterfactuals while managing effectively this trade-off, and guarantees the user a minimal robustness.
arXiv Detail & Related papers (2023-04-24T09:00:31Z) - Explaining Recommendation System Using Counterfactual Textual
Explanations [4.318555434063274]
It is found that if end-users understand the reason for the production of some output, it is easier to trust the system.
One method for producing a more explainable output is using counterfactual reasoning.
arXiv Detail & Related papers (2023-03-14T06:45:28Z) - Active Learning for Abstractive Text Summarization [50.79416783266641]
We propose the first effective query strategy for Active Learning in abstractive text summarization.
We show that using our strategy in AL annotation helps to improve the model performance in terms of ROUGE and consistency scores.
arXiv Detail & Related papers (2023-01-09T10:33:14Z) - Complementary Explanations for Effective In-Context Learning [77.83124315634386]
Large language models (LLMs) have exhibited remarkable capabilities in learning from explanations in prompts.
This work aims to better understand the mechanisms by which explanations are used for in-context learning.
arXiv Detail & Related papers (2022-11-25T04:40:47Z) - MACE: An Efficient Model-Agnostic Framework for Counterfactual
Explanation [132.77005365032468]
We propose a novel framework of Model-Agnostic Counterfactual Explanation (MACE)
In our MACE approach, we propose a novel RL-based method for finding good counterfactual examples and a gradient-less descent method for improving proximity.
Experiments on public datasets validate the effectiveness with better validity, sparsity and proximity.
arXiv Detail & Related papers (2022-05-31T04:57:06Z) - Synthesizing explainable counterfactual policies for algorithmic
recourse with program synthesis [18.485744170172545]
We learn a program that outputs a sequence of explainable counterfactual actions given a user description and a causal graph.
An experimental evaluation on synthetic and real-world datasets shows how our approach generates effective interventions.
arXiv Detail & Related papers (2022-01-18T17:16:45Z) - Search Methods for Sufficient, Socially-Aligned Feature Importance
Explanations with In-Distribution Counterfactuals [72.00815192668193]
Feature importance (FI) estimates are a popular form of explanation, and they are commonly created and evaluated by computing the change in model confidence caused by removing certain input features at test time.
We study several under-explored dimensions of FI-based explanations, providing conceptual and empirical improvements for this form of explanation.
arXiv Detail & Related papers (2021-06-01T20:36:48Z) - Towards Unifying Feature Attribution and Counterfactual Explanations:
Different Means to the Same End [17.226134854746267]
We present a method to generate feature attribution explanations from a set of counterfactual examples.
We show how counterfactual examples can be used to evaluate the goodness of an attribution-based explanation in terms of its necessity and sufficiency.
arXiv Detail & Related papers (2020-11-10T05:41:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.