SLISEMAP: Explainable Dimensionality Reduction
- URL: http://arxiv.org/abs/2201.04455v1
- Date: Wed, 12 Jan 2022 13:06:21 GMT
- Title: SLISEMAP: Explainable Dimensionality Reduction
- Authors: Anton Bj\"orklund, Jarmo M\"akel\"a, Kai Puolam\"aki
- Abstract summary: Existing explanation methods for black-box supervised learning models generally work by building local models that explain the models behaviour for a particular data item.
We propose a new manifold visualization method, SLISEMAP, that finds local explanations for all of the data items and builds a two-dimensional visualization of model space.
We show that SLISEMAP provides fast and stable visualizations that can be used to explain and understand black box regression and classification models.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Existing explanation methods for black-box supervised learning models
generally work by building local models that explain the models behaviour for a
particular data item. It is possible to make global explanations, but the
explanations may have low fidelity for complex models. Most of the prior work
on explainable models has been focused on classification problems, with less
attention on regression.
We propose a new manifold visualization method, SLISEMAP, that at the same
time finds local explanations for all of the data items and builds a
two-dimensional visualization of model space such that the data items explained
by the same model are projected nearby. We provide an open source
implementation of our methods, implemented by using GPU-optimized PyTorch
library. SLISEMAP works both on classification and regression models.
We compare SLISEMAP to most popular dimensionality reduction methods and some
local explanation methods. We provide mathematical derivation of our problem
and show that SLISEMAP provides fast and stable visualizations that can be used
to explain and understand black box regression and classification models.
Related papers
- Interpretability in Symbolic Regression: a benchmark of Explanatory Methods using the Feynman data set [0.0]
Interpretability of machine learning models plays a role as important as the model accuracy.
This paper proposes a benchmark scheme to evaluate explanatory methods to explain regression models.
Results have shown that Symbolic Regression models can be an interesting alternative to white-box and black-box models.
arXiv Detail & Related papers (2024-04-08T23:46:59Z) - Learning with Explanation Constraints [91.23736536228485]
We provide a learning theoretic framework to analyze how explanations can improve the learning of our models.
We demonstrate the benefits of our approach over a large array of synthetic and real-world experiments.
arXiv Detail & Related papers (2023-03-25T15:06:47Z) - Streamlining models with explanations in the learning loop [0.0]
Several explainable AI methods allow a Machine Learning user to get insights on the classification process of a black-box model.
We exploit this information to design a feature engineering phase, where we combine explanations with feature values.
arXiv Detail & Related papers (2023-02-15T16:08:32Z) - The Manifold Hypothesis for Gradient-Based Explanations [55.01671263121624]
gradient-based explanation algorithms provide perceptually-aligned explanations.
We show that the more a feature attribution is aligned with the tangent space of the data, the more perceptually-aligned it tends to be.
We suggest that explanation algorithms should actively strive to align their explanations with the data manifold.
arXiv Detail & Related papers (2022-06-15T08:49:24Z) - GRAPHSHAP: Explaining Identity-Aware Graph Classifiers Through the
Language of Motifs [11.453325862543094]
GRAPHSHAP is able to provide motif-based explanations for identity-aware graph classifiers.
We show how a simple kernel can efficiently approximate explanation scores, thus allowing GRAPHSHAP to scale on scenarios with a large explanation space.
Our experiments highlight how the classification provided by a black-box model can be effectively explained by few connectomics patterns.
arXiv Detail & Related papers (2022-02-17T18:29:30Z) - Beyond Trivial Counterfactual Explanations with Diverse Valuable
Explanations [64.85696493596821]
In computer vision applications, generative counterfactual methods indicate how to perturb a model's input to change its prediction.
We propose a counterfactual method that learns a perturbation in a disentangled latent space that is constrained using a diversity-enforcing loss.
Our model improves the success rate of producing high-quality valuable explanations when compared to previous state-of-the-art methods.
arXiv Detail & Related papers (2021-03-18T12:57:34Z) - Visualising Deep Network's Time-Series Representations [93.73198973454944]
Despite the popularisation of machine learning models, more often than not they still operate as black boxes with no insight into what is happening inside the model.
In this paper, a method that addresses that issue is proposed, with a focus on visualising multi-dimensional time-series data.
Experiments on a high-frequency stock market dataset show that the method provides fast and discernible visualisations.
arXiv Detail & Related papers (2021-03-12T09:53:34Z) - Leakage-Adjusted Simulatability: Can Models Generate Non-Trivial
Explanations of Their Behavior in Natural Language? [86.60613602337246]
We introduce a leakage-adjusted simulatability (LAS) metric for evaluating NL explanations.
LAS measures how well explanations help an observer predict a model's output, while controlling for how explanations can directly leak the output.
We frame explanation generation as a multi-agent game and optimize explanations for simulatability while penalizing label leakage.
arXiv Detail & Related papers (2020-10-08T16:59:07Z) - MeLIME: Meaningful Local Explanation for Machine Learning Models [2.819725769698229]
We show that our approach, MeLIME, produces more meaningful explanations compared to other techniques over different ML models.
MeLIME generalizes the LIME method, allowing more flexible perturbation sampling and the use of different local interpretable models.
arXiv Detail & Related papers (2020-09-12T16:06:58Z) - Deducing neighborhoods of classes from a fitted model [68.8204255655161]
In this article a new kind of interpretable machine learning method is presented.
It can help to understand the partitioning of the feature space into predicted classes in a classification model using quantile shifts.
Basically, real data points (or specific points of interest) are used and the changes of the prediction after slightly raising or decreasing specific features are observed.
arXiv Detail & Related papers (2020-09-11T16:35:53Z) - Explainable Deep Modeling of Tabular Data using TableGraphNet [1.376408511310322]
We propose a new architecture that produces explainable predictions in the form of additive feature attributions.
We show that our explainable model attains the same level of performance as black box models.
arXiv Detail & Related papers (2020-02-12T20:02:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.