Transforming Feature Space to Interpret Machine Learning Models
- URL: http://arxiv.org/abs/2104.04295v1
- Date: Fri, 9 Apr 2021 10:48:11 GMT
- Title: Transforming Feature Space to Interpret Machine Learning Models
- Authors: Alexander Brenning
- Abstract summary: This contribution proposes a novel approach that interprets machine-learning models through the lens of feature space transformations.
It can be used to enhance unconditional as well as conditional post-hoc diagnostic tools.
A case study on remote-sensing landcover classification with 46 features is used to demonstrate the potential of the proposed approach.
- Score: 91.62936410696409
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Model-agnostic tools for interpreting machine-learning models struggle to
summarize the joint effects of strongly dependent features in high-dimensional
feature spaces, which play an important role in pattern recognition, for
example in remote sensing of landcover. This contribution proposes a novel
approach that interprets machine-learning models through the lens of feature
space transformations. It can be used to enhance unconditional as well as
conditional post-hoc diagnostic tools including partial dependence plots,
accumulated local effects plots, or permutation feature importance assessments.
While the approach can also be applied to nonlinear transformations, we focus
on linear ones, including principal component analysis (PCA) and a partial
orthogonalization technique. Structured PCA and diagnostics along paths offer
opportunities for representing domain knowledge. The new approach is
implemented in the R package `wiml`, which can be combined with existing
explainable machine-learning packages. A case study on remote-sensing landcover
classification with 46 features is used to demonstrate the potential of the
proposed approach for model interpretation by domain experts.
Related papers
- Principal Orthogonal Latent Components Analysis (POLCA Net) [0.27309692684728604]
representation learning aims to learn features that are more useful and relevant for tasks such as classification, prediction, and clustering.
We introduce Principal Orthogonal Latent Components Analysis Network (POLCA Net), an approach to mimic and extend PCA and LDA capabilities to non-linear domains.
arXiv Detail & Related papers (2024-10-09T14:04:31Z) - Prospector Heads: Generalized Feature Attribution for Large Models & Data [82.02696069543454]
We introduce prospector heads, an efficient and interpretable alternative to explanation-based attribution methods.
We demonstrate how prospector heads enable improved interpretation and discovery of class-specific patterns in input data.
arXiv Detail & Related papers (2024-02-18T23:01:28Z) - Relational Local Explanations [11.679389861042]
We develop a novel model-agnostic and permutation-based feature attribution algorithm based on relational analysis between input variables.
We are able to gain a broader insight into machine learning model decisions and data.
arXiv Detail & Related papers (2022-12-23T14:46:23Z) - Towards a mathematical understanding of learning from few examples with
nonlinear feature maps [68.8204255655161]
We consider the problem of data classification where the training set consists of just a few data points.
We reveal key relationships between the geometry of an AI model's feature space, the structure of the underlying data distributions, and the model's generalisation capabilities.
arXiv Detail & Related papers (2022-11-07T14:52:58Z) - MACE: An Efficient Model-Agnostic Framework for Counterfactual
Explanation [132.77005365032468]
We propose a novel framework of Model-Agnostic Counterfactual Explanation (MACE)
In our MACE approach, we propose a novel RL-based method for finding good counterfactual examples and a gradient-less descent method for improving proximity.
Experiments on public datasets validate the effectiveness with better validity, sparsity and proximity.
arXiv Detail & Related papers (2022-05-31T04:57:06Z) - Spatial machine-learning model diagnostics: a model-agnostic
distance-based approach [91.62936410696409]
This contribution proposes spatial prediction error profiles (SPEPs) and spatial variable importance profiles (SVIPs) as novel model-agnostic assessment and interpretation tools.
The SPEPs and SVIPs of geostatistical methods, linear models, random forest, and hybrid algorithms show striking differences and also relevant similarities.
The novel diagnostic tools enrich the toolkit of spatial data science, and may improve ML model interpretation, selection, and design.
arXiv Detail & Related papers (2021-11-13T01:50:36Z) - Grouped Feature Importance and Combined Features Effect Plot [2.15867006052733]
Interpretable machine learning has become a very active area of research due to the rising popularity of machine learning algorithms.
We provide a comprehensive overview of how existing model-agnostic techniques can be defined for feature groups to assess the grouped feature importance.
We introduce the combined features effect plot, which is a technique to visualize the effect of a group of features based on a sparse, interpretable linear combination of features.
arXiv Detail & Related papers (2021-04-23T16:27:38Z) - Controlling for sparsity in sparse factor analysis models: adaptive
latent feature sharing for piecewise linear dimensionality reduction [2.896192909215469]
We propose a simple and tractable parametric feature allocation model which can address key limitations of current latent feature decomposition techniques.
We derive a novel adaptive Factor analysis (aFA), as well as, an adaptive probabilistic principle component analysis (aPPCA) capable of flexible structure discovery and dimensionality reduction.
We show that aPPCA and aFA can infer interpretable high level features both when applied on raw MNIST and when applied for interpreting autoencoder features.
arXiv Detail & Related papers (2020-06-22T16:09:11Z) - A Trainable Optimal Transport Embedding for Feature Aggregation and its
Relationship to Attention [96.77554122595578]
We introduce a parametrized representation of fixed size, which embeds and then aggregates elements from a given input set according to the optimal transport plan between the set and a trainable reference.
Our approach scales to large datasets and allows end-to-end training of the reference, while also providing a simple unsupervised learning mechanism with small computational cost.
arXiv Detail & Related papers (2020-06-22T08:35:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.