Consistent Sufficient Explanations and Minimal Local Rules for
explaining regression and classification models
- URL: http://arxiv.org/abs/2111.04658v1
- Date: Mon, 8 Nov 2021 17:27:52 GMT
- Title: Consistent Sufficient Explanations and Minimal Local Rules for
explaining regression and classification models
- Authors: Salim I. Amoukou and Nicolas J.B Brunel
- Abstract summary: We extend the notion of probabilistic Sufficient Explanations (P-SE)
The crux of P-SE is to compute the conditional probability of maintaining the same prediction.
We deal with non-binary features, without learning the distribution of $X$ nor having the model for making predictions.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: To explain the decision of any model, we extend the notion of probabilistic
Sufficient Explanations (P-SE). For each instance, this approach selects the
minimal subset of features that is sufficient to yield the same prediction with
high probability, while removing other features. The crux of P-SE is to compute
the conditional probability of maintaining the same prediction. Therefore, we
introduce an accurate and fast estimator of this probability via random Forests
for any data $(\boldsymbol{X}, Y)$ and show its efficiency through a
theoretical analysis of its consistency. As a consequence, we extend the P-SE
to regression problems. In addition, we deal with non-binary features, without
learning the distribution of $X$ nor having the model for making predictions.
Finally, we introduce local rule-based explanations for
regression/classification based on the P-SE and compare our approaches w.r.t
other explainable AI methods. These methods are publicly available as a Python
package at \url{www.github.com/salimamoukou/acv00}.
Related papers
- Rejection via Learning Density Ratios [50.91522897152437]
Classification with rejection emerges as a learning paradigm which allows models to abstain from making predictions.
We propose a different distributional perspective, where we seek to find an idealized data distribution which maximizes a pretrained model's performance.
Our framework is tested empirically over clean and noisy datasets.
arXiv Detail & Related papers (2024-05-29T01:32:17Z) - Online non-parametric likelihood-ratio estimation by Pearson-divergence
functional minimization [55.98760097296213]
We introduce a new framework for online non-parametric LRE (OLRE) for the setting where pairs of iid observations $(x_t sim p, x'_t sim q)$ are observed over time.
We provide theoretical guarantees for the performance of the OLRE method along with empirical validation in synthetic experiments.
arXiv Detail & Related papers (2023-11-03T13:20:11Z) - Calibrated Explanations for Regression [1.2058600649065616]
Calibrated Explanations for regression provides fast, reliable, stable, and robust explanations.
Calibrated Explanations for probabilistic regression provides an entirely new way of creating explanations.
An implementation in Python is freely available on GitHub and for installation using both pip and conda.
arXiv Detail & Related papers (2023-08-30T18:06:57Z) - Performative Prediction with Neural Networks [24.880495520422]
performative prediction is a framework for learning models that influence the data they intend to predict.
Standard convergence results for finding a performatively stable classifier with the method of repeated risk minimization assume that the data distribution is Lipschitz continuous to the model's parameters.
In this work, we instead assume that the data distribution is Lipschitz continuous with respect to the model's predictions, a more natural assumption for performative systems.
arXiv Detail & Related papers (2023-04-14T01:12:48Z) - Ensemble Multi-Quantiles: Adaptively Flexible Distribution Prediction
for Uncertainty Quantification [4.728311759896569]
We propose a novel, succinct, and effective approach for distribution prediction to quantify uncertainty in machine learning.
It incorporates adaptively flexible distribution prediction of $mathbbP(mathbfy|mathbfX=x)$ in regression tasks.
On extensive regression tasks from UCI datasets, we show that EMQ achieves state-of-the-art performance.
arXiv Detail & Related papers (2022-11-26T11:45:32Z) - Exact and Approximate Conformal Inference for Multi-Output Regression [0.0]
Conformal inference is used in machine learning to quantify uncertainty associated with predictions.
In this paper, we explore multi-output regression, delivering exact derivations of conformal inference $p$-values.
We also provide both theoretical and empirical evidence of the effectiveness of these methods using both real-world and simulated data.
arXiv Detail & Related papers (2022-10-31T15:41:13Z) - An Imprecise SHAP as a Tool for Explaining the Class Probability
Distributions under Limited Training Data [5.8010446129208155]
An imprecise SHAP is proposed for cases when the class probability distributions are imprecise and represented by sets of distributions.
The first idea behind the imprecise SHAP is a new approach for computing the marginal contribution of a feature.
The second idea is an attempt to consider a general approach to calculating and reducing interval-valued Shapley values.
arXiv Detail & Related papers (2021-06-16T20:30:26Z) - Multivariate Probabilistic Regression with Natural Gradient Boosting [63.58097881421937]
We propose a Natural Gradient Boosting (NGBoost) approach based on nonparametrically modeling the conditional parameters of the multivariate predictive distribution.
Our method is robust, works out-of-the-box without extensive tuning, is modular with respect to the assumed target distribution, and performs competitively in comparison to existing approaches.
arXiv Detail & Related papers (2021-06-07T17:44:49Z) - Probabilistic Gradient Boosting Machines for Large-Scale Probabilistic
Regression [51.770998056563094]
Probabilistic Gradient Boosting Machines (PGBM) is a method to create probabilistic predictions with a single ensemble of decision trees.
We empirically demonstrate the advantages of PGBM compared to existing state-of-the-art methods.
arXiv Detail & Related papers (2021-06-03T08:32:13Z) - Learning Probabilistic Ordinal Embeddings for Uncertainty-Aware
Regression [91.3373131262391]
Uncertainty is the only certainty there is.
Traditionally, the direct regression formulation is considered and the uncertainty is modeled by modifying the output space to a certain family of probabilistic distributions.
How to model the uncertainty within the present-day technologies for regression remains an open issue.
arXiv Detail & Related papers (2021-03-25T06:56:09Z) - Decision-Making with Auto-Encoding Variational Bayes [71.44735417472043]
We show that a posterior approximation distinct from the variational distribution should be used for making decisions.
Motivated by these theoretical results, we propose learning several approximate proposals for the best model.
In addition to toy examples, we present a full-fledged case study of single-cell RNA sequencing.
arXiv Detail & Related papers (2020-02-17T19:23:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.