Counterfactual Explanation via Search in Gaussian Mixture Distributed
Latent Space
- URL: http://arxiv.org/abs/2307.13390v3
- Date: Tue, 21 Nov 2023 19:11:34 GMT
- Title: Counterfactual Explanation via Search in Gaussian Mixture Distributed
Latent Space
- Authors: Xuan Zhao, Klaus Broelemann, Gjergji Kasneci
- Abstract summary: Counterfactual Explanations (CEs) are an important tool in Algorithmic Recourse for addressing two questions.
guiding the user's interaction with AI systems by proposing easy-to-understand explanations is essential for the trustworthy adoption and long-term acceptance of AI systems.
We introduce a new method to generate CEs for a pre-trained binary classifier by first shaping the latent space of an autoencoder to be a mixture of Gaussian distributions.
- Score: 19.312306559210125
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Counterfactual Explanations (CEs) are an important tool in Algorithmic
Recourse for addressing two questions: 1. What are the crucial factors that led
to an automated prediction/decision? 2. How can these factors be changed to
achieve a more favorable outcome from a user's perspective? Thus, guiding the
user's interaction with AI systems by proposing easy-to-understand explanations
and easy-to-attain feasible changes is essential for the trustworthy adoption
and long-term acceptance of AI systems. In the literature, various methods have
been proposed to generate CEs, and different quality measures have been
suggested to evaluate these methods. However, the generation of CEs is usually
computationally expensive, and the resulting suggestions are unrealistic and
thus non-actionable. In this paper, we introduce a new method to generate CEs
for a pre-trained binary classifier by first shaping the latent space of an
autoencoder to be a mixture of Gaussian distributions. CEs are then generated
in latent space by linear interpolation between the query sample and the
centroid of the target class. We show that our method maintains the
characteristics of the input sample during the counterfactual search. In
various experiments, we show that the proposed method is competitive based on
different quality measures on image and tabular datasets -- efficiently returns
results that are closer to the original data manifold compared to three
state-of-the-art methods, which are essential for realistic high-dimensional
machine learning applications.
Related papers
- Interval Abstractions for Robust Counterfactual Explanations [15.954944873701503]
We propose a novel interval abstraction technique for parametric machine learning models.
We obtain provable robustness guarantees of CEs under the possibly infinite set of plausible model changes $Delta$.
arXiv Detail & Related papers (2024-04-21T18:24:34Z) - GCC: Generative Calibration Clustering [55.44944397168619]
We propose a novel Generative Clustering (GCC) method to incorporate feature learning and augmentation into clustering procedure.
First, we develop a discrimirative feature alignment mechanism to discover intrinsic relationship across real and generated samples.
Second, we design a self-supervised metric learning to generate more reliable cluster assignment.
arXiv Detail & Related papers (2024-04-14T01:51:11Z) - Introducing User Feedback-based Counterfactual Explanations (UFCE) [49.1574468325115]
Counterfactual explanations (CEs) have emerged as a viable solution for generating comprehensible explanations in XAI.
UFCE allows for the inclusion of user constraints to determine the smallest modifications in the subset of actionable features.
UFCE outperforms two well-known CE methods in terms of textitproximity, textitsparsity, and textitfeasibility.
arXiv Detail & Related papers (2024-02-26T20:09:44Z) - A Weighted K-Center Algorithm for Data Subset Selection [70.49696246526199]
Subset selection is a fundamental problem that can play a key role in identifying smaller portions of the training data.
We develop a novel factor 3-approximation algorithm to compute subsets based on the weighted sum of both k-center and uncertainty sampling objective functions.
arXiv Detail & Related papers (2023-12-17T04:41:07Z) - Counterfactual Explanation for Regression via Disentanglement in Latent
Space [19.312306559210125]
We introduce a novel method to generate Counterfactual Explanations (CEs) for a pre-trained regressor.
We show that our method maintains the characteristics of the query sample during the counterfactual search.
Our code will be made available as an open-source package upon the publication of this work.
arXiv Detail & Related papers (2023-11-14T15:08:14Z) - Provably Robust and Plausible Counterfactual Explanations for Neural Networks via Robust Optimisation [19.065904250532995]
We propose Provably RObust and PLAusible Counterfactual Explanations (PROPLACE)
We formulate an iterative algorithm to compute provably robust CEs and prove its convergence, soundness and completeness.
We show that PROPLACE achieves state-of-the-art performances against metrics on three evaluation aspects.
arXiv Detail & Related papers (2023-09-22T00:12:09Z) - Diffusion-based Visual Counterfactual Explanations -- Towards Systematic
Quantitative Evaluation [64.0476282000118]
Latest methods for visual counterfactual explanations (VCE) harness the power of deep generative models to synthesize new examples of high-dimensional images of impressive quality.
It is currently difficult to compare the performance of these VCE methods as the evaluation procedures largely vary and often boil down to visual inspection of individual examples and small scale user studies.
We propose a framework for systematic, quantitative evaluation of the VCE methods and a minimal set of metrics to be used.
arXiv Detail & Related papers (2023-08-11T12:22:37Z) - Finding Regions of Counterfactual Explanations via Robust Optimization [0.0]
A counterfactual explanation (CE) is a minimal perturbed data point for which the decision of the model changes.
Most of the existing methods can only provide one CE, which may not be achievable for the user.
We derive an iterative method to calculate robust CEs that remain valid even after the features are slightly perturbed.
arXiv Detail & Related papers (2023-01-26T14:06:26Z) - Deep Active Ensemble Sampling For Image Classification [8.31483061185317]
Active learning frameworks aim to reduce the cost of data annotation by actively requesting the labeling for the most informative data points.
Some proposed approaches include uncertainty-based techniques, geometric methods, implicit combination of uncertainty-based and geometric approaches.
We present an innovative integration of recent progress in both uncertainty-based and geometric frameworks to enable an efficient exploration/exploitation trade-off in sample selection strategy.
Our framework provides two advantages: (1) accurate posterior estimation, and (2) tune-able trade-off between computational overhead and higher accuracy.
arXiv Detail & Related papers (2022-10-11T20:20:20Z) - Large-Scale Sequential Learning for Recommender and Engineering Systems [91.3755431537592]
In this thesis, we focus on the design of an automatic algorithms that provide personalized ranking by adapting to the current conditions.
For the former, we propose novel algorithm called SAROS that take into account both kinds of feedback for learning over the sequence of interactions.
The proposed idea of taking into account the neighbour lines shows statistically significant results in comparison with the initial approach for faults detection in power grid.
arXiv Detail & Related papers (2022-05-13T21:09:41Z) - An Online Method for A Class of Distributionally Robust Optimization
with Non-Convex Objectives [54.29001037565384]
We propose a practical online method for solving a class of online distributionally robust optimization (DRO) problems.
Our studies demonstrate important applications in machine learning for improving the robustness of networks.
arXiv Detail & Related papers (2020-06-17T20:19:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.