Local Universal Explainer (LUX) -- a rule-based explainer with factual, counterfactual and visual explanations
- URL: http://arxiv.org/abs/2310.14894v3
- Date: Mon, 9 Sep 2024 07:07:11 GMT
- Title: Local Universal Explainer (LUX) -- a rule-based explainer with factual, counterfactual and visual explanations
- Authors: Szymon Bobek, Grzegorz J. Nalepa,
- Abstract summary: Local Universal Explainer (LUX) is a rule-based explainer that can generate factual, counterfactual and visual explanations.
It is based on a modified version of decision tree algorithms that allows for oblique splits and integration with feature importance XAI methods such as SHAP.
We tested our method on real and synthetic datasets and compared it with state-of-the-art rule-based explainers such as LORE, EXPLAN and Anchor.
- Score: 7.673339435080445
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Explainable artificial intelligence (XAI) is one of the most intensively developed area of AI in recent years. It is also one of the most fragmented with multiple methods that focus on different aspects of explanations. This makes difficult to obtain the full spectrum of explanation at once in a compact and consistent way. To address this issue, we present Local Universal Explainer (LUX), which is a rule-based explainer that can generate factual, counterfactual and visual explanations. It is based on a modified version of decision tree algorithms that allows for oblique splits and integration with feature importance XAI methods such as SHAP. It limits the use data generation in opposite to other algorithms, but is focused on selecting local concepts in a form of high-density clusters of real data that have the highest impact on forming the decision boundary of the explained model and generating artificial samples with novel SHAP-guided sampling algorithm. We tested our method on real and synthetic datasets and compared it with state-of-the-art rule-based explainers such as LORE, EXPLAN and Anchor. Our method outperforms the existing approaches in terms of simplicity, fidelity, representativeness, and consistency.
Related papers
- Info-CELS: Informative Saliency Map Guided Counterfactual Explanation [1.25828876338076]
A novel counterfactual explanation model, CELS, has been introduced.
CELS learns a saliency map for the interest of an instance and generates a counterfactual explanation guided by the learned saliency map.
We present an enhanced approach that builds upon CELS.
arXiv Detail & Related papers (2024-10-27T18:12:02Z) - Global Human-guided Counterfactual Explanations for Molecular Properties via Reinforcement Learning [49.095065258759895]
We develop a novel global explanation model RLHEX for molecular property prediction.
It aligns the counterfactual explanations with human-defined principles, making the explanations more interpretable and easy for experts to evaluate.
The global explanations produced by RLHEX cover 4.12% more input graphs and reduce the distance between the counterfactual explanation set and the input set by 0.47% on average across three molecular datasets.
arXiv Detail & Related papers (2024-06-19T22:16:40Z) - CAGE: Causality-Aware Shapley Value for Global Explanations [4.017708359820078]
One way to explain AI models is to elucidate the predictive importance of the input features for the AI model.
Inspired by cooperative game theory, Shapley values offer a convenient way for quantifying the feature importance as explanations.
In particular, we introduce a novel sampling procedure for out-coalition features that respects the causal relations of the input features.
arXiv Detail & Related papers (2024-04-17T09:43:54Z) - Fidelity of Ensemble Aggregation for Saliency Map Explanations using
Bayesian Optimization Techniques [0.0]
We present and compare different pixel-based aggregation schemes with the goal of generating a new explanation.
We incorporate the variance between the individual explanations into the aggregation process.
We also analyze the effect of multiple normalization techniques on ensemble aggregation.
arXiv Detail & Related papers (2022-07-04T16:34:12Z) - Principled Knowledge Extrapolation with GANs [92.62635018136476]
We study counterfactual synthesis from a new perspective of knowledge extrapolation.
We show that an adversarial game with a closed-form discriminator can be used to address the knowledge extrapolation problem.
Our method enjoys both elegant theoretical guarantees and superior performance in many scenarios.
arXiv Detail & Related papers (2022-05-21T08:39:42Z) - Clustered Federated Learning via Generalized Total Variation
Minimization [83.26141667853057]
We study optimization methods to train local (or personalized) models for local datasets with a decentralized network structure.
Our main conceptual contribution is to formulate federated learning as total variation minimization (GTV)
Our main algorithmic contribution is a fully decentralized federated learning algorithm.
arXiv Detail & Related papers (2021-05-26T18:07:19Z) - Deep Magnification-Flexible Upsampling over 3D Point Clouds [103.09504572409449]
We propose a novel end-to-end learning-based framework to generate dense point clouds.
We first formulate the problem explicitly, which boils down to determining the weights and high-order approximation errors.
Then, we design a lightweight neural network to adaptively learn unified and sorted weights as well as the high-order refinements.
arXiv Detail & Related papers (2020-11-25T14:00:18Z) - Explaining Convolutional Neural Networks through Attribution-Based Input
Sampling and Block-Wise Feature Aggregation [22.688772441351308]
Methods based on class activation mapping and randomized input sampling have gained great popularity.
However, the attribution methods provide lower resolution and blurry explanation maps that limit their explanation power.
In this work, we collect visualization maps from multiple layers of the model based on an attribution-based input sampling technique.
We also propose a layer selection strategy that applies to the whole family of CNN-based models.
arXiv Detail & Related papers (2020-10-01T20:27:30Z) - FedPD: A Federated Learning Framework with Optimal Rates and Adaptivity
to Non-IID Data [59.50904660420082]
Federated Learning (FL) has become a popular paradigm for learning from distributed data.
To effectively utilize data at different devices without moving them to the cloud, algorithms such as the Federated Averaging (FedAvg) have adopted a "computation then aggregation" (CTA) model.
arXiv Detail & Related papers (2020-05-22T23:07:42Z) - The data-driven physical-based equations discovery using evolutionary
approach [77.34726150561087]
We describe the algorithm for the mathematical equations discovery from the given observations data.
The algorithm combines genetic programming with the sparse regression.
It could be used for governing analytical equation discovery as well as for partial differential equations (PDE) discovery.
arXiv Detail & Related papers (2020-04-03T17:21:57Z) - Ontology-based Interpretable Machine Learning for Textual Data [35.01650633374998]
We introduce a novel interpreting framework that learns an interpretable model based on sampling technique to explain prediction models.
To narrow down the search space for explanations, we design a learnable anchor algorithm.
A set of regulations is further introduced, regarding combining learned interpretable representations with anchors to generate comprehensible explanations.
arXiv Detail & Related papers (2020-04-01T02:51:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.