Generating robust counterfactual explanations
- URL: http://arxiv.org/abs/2304.12943v1
- Date: Mon, 24 Apr 2023 09:00:31 GMT
- Title: Generating robust counterfactual explanations
- Authors: Victor Guyomard, Fran\c{c}oise Fessant, Thomas Guyet, Tassadit Bouadi,
Alexandre Termier
- Abstract summary: The quality of a counterfactual depends on several criteria: realism, actionability, validity, robustness, etc.
In this paper, we are interested in the notion of robustness of a counterfactual. More precisely, we focus on robustness to counterfactual input changes.
We propose a new framework, CROCO, that generates robust counterfactuals while managing effectively this trade-off, and guarantees the user a minimal robustness.
- Score: 60.32214822437734
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Counterfactual explanations have become a mainstay of the XAI field. This
particularly intuitive statement allows the user to understand what small but
necessary changes would have to be made to a given situation in order to change
a model prediction. The quality of a counterfactual depends on several
criteria: realism, actionability, validity, robustness, etc. In this paper, we
are interested in the notion of robustness of a counterfactual. More precisely,
we focus on robustness to counterfactual input changes. This form of robustness
is particularly challenging as it involves a trade-off between the robustness
of the counterfactual and the proximity with the example to explain. We propose
a new framework, CROCO, that generates robust counterfactuals while managing
effectively this trade-off, and guarantees the user a minimal robustness. An
empirical evaluation on tabular datasets confirms the relevance and
effectiveness of our approach.
Related papers
- Rigorous Probabilistic Guarantees for Robust Counterfactual Explanations [80.86128012438834]
We show for the first time that computing the robustness of counterfactuals with respect to plausible model shifts is NP-complete.
We propose a novel probabilistic approach which is able to provide tight estimates of robustness with strong guarantees.
arXiv Detail & Related papers (2024-07-10T09:13:11Z) - Promoting Counterfactual Robustness through Diversity [10.223545393731115]
Counterfactual explainers may lack robustness in the sense that a minor change in the input can cause a major change in the explanation.
We propose an approximation algorithm that uses a diversity criterion to select a feasible number of most relevant explanations.
arXiv Detail & Related papers (2023-12-11T17:49:25Z) - Flexible and Robust Counterfactual Explanations with Minimal Satisfiable
Perturbations [56.941276017696076]
We propose a conceptually simple yet effective solution named Counterfactual Explanations with Minimal Satisfiable Perturbations (CEMSP)
CEMSP constrains changing values of abnormal features with the help of their semantically meaningful normal ranges.
Compared to existing methods, we conduct comprehensive experiments on both synthetic and real-world datasets to demonstrate that our method provides more robust explanations while preserving flexibility.
arXiv Detail & Related papers (2023-09-09T04:05:56Z) - Feature-based Learning for Diverse and Privacy-Preserving Counterfactual
Explanations [46.89706747651661]
Interpretable machine learning seeks to understand the reasoning process of complex black-box systems.
One flourishing approach is through counterfactual explanations, which provide suggestions on what a user can do to alter an outcome.
arXiv Detail & Related papers (2022-09-27T15:09:13Z) - Exploring the Trade-off between Plausibility, Change Intensity and
Adversarial Power in Counterfactual Explanations using Multi-objective
Optimization [73.89239820192894]
We argue that automated counterfactual generation should regard several aspects of the produced adversarial instances.
We present a novel framework for the generation of counterfactual examples.
arXiv Detail & Related papers (2022-05-20T15:02:53Z) - Robustness and Accuracy Could Be Reconcilable by (Proper) Definition [109.62614226793833]
The trade-off between robustness and accuracy has been widely studied in the adversarial literature.
We find that it may stem from the improperly defined robust error, which imposes an inductive bias of local invariance.
By definition, SCORE facilitates the reconciliation between robustness and accuracy, while still handling the worst-case uncertainty.
arXiv Detail & Related papers (2022-02-21T10:36:09Z) - Reliable Post hoc Explanations: Modeling Uncertainty in Explainability [44.9824285459365]
Black box explanations are increasingly being employed to establish model credibility in high-stakes settings.
prior work demonstrates that explanations generated by state-of-the-art techniques are inconsistent, unstable, and provide very little insight into their correctness and reliability.
We develop a novel Bayesian framework for generating local explanations along with their associated uncertainty.
arXiv Detail & Related papers (2020-08-11T22:52:21Z) - Unifying Model Explainability and Robustness via Machine-Checkable
Concepts [33.88198813484126]
We propose a robustness-assessment framework, at the core of which is the idea of using machine-checkable concepts.
Our framework defines a large number of concepts that the explanations could be based on and performs the explanation-conformity check at test time to assess prediction robustness.
Experiments on real-world datasets and human surveys show that our framework is able to enhance prediction robustness significantly.
arXiv Detail & Related papers (2020-07-01T05:21:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.