MCCE: Monte Carlo sampling of realistic counterfactual explanations
- URL: http://arxiv.org/abs/2111.09790v2
- Date: Thu, 25 Jan 2024 04:39:53 GMT
- Title: MCCE: Monte Carlo sampling of realistic counterfactual explanations
- Authors: Annabelle Redelmeier, Martin Jullum, Kjersti Aas, Anders L{\o}land
- Abstract summary: MCCE is a novel on-manifold, actionable and valid counterfactual explanation method.
It generates on-manifold, actionable and valid counterfactuals by modeling the joint distribution of the mutable features.
We compare MCCE with a range of state-of-the-art on-manifold counterfactual methods using four well-known data sets.
- Score: 2.156170153103442
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We introduce MCCE: Monte Carlo sampling of valid and realistic Counterfactual
Explanations for tabular data, a novel counterfactual explanation method that
generates on-manifold, actionable and valid counterfactuals by modeling the
joint distribution of the mutable features given the immutable features and the
decision. Unlike other on-manifold methods that tend to rely on variational
autoencoders and have strict prediction model and data requirements, MCCE
handles any type of prediction model and categorical features with more than
two levels. MCCE first models the joint distribution of the features and the
decision with an autoregressive generative model where the conditionals are
estimated using decision trees. Then, it samples a large set of observations
from this model, and finally, it removes the samples that do not obey certain
criteria. We compare MCCE with a range of state-of-the-art on-manifold
counterfactual methods using four well-known data sets and show that MCCE
outperforms these methods on all common performance metrics and speed. In
particular, including the decision in the modeling process improves the
efficiency of the method substantially.
Related papers
- Influence Functions for Scalable Data Attribution in Diffusion Models [52.92223039302037]
Diffusion models have led to significant advancements in generative modelling.
Yet their widespread adoption poses challenges regarding data attribution and interpretability.
In this paper, we aim to help address such challenges by developing an textitinfluence functions framework.
arXiv Detail & Related papers (2024-10-17T17:59:02Z) - Robust Inference of Dynamic Covariance Using Wishart Processes and Sequential Monte Carlo [2.6347238599620115]
We introduce a Sequential Monte Carlo (SMC) sampler for the Wishart process.
We show that SMC sampling results in the most robust estimates and out-of-sample predictions of dynamic covariance.
We demonstrate the practical applicability of our proposed approach on a dataset of clinical depression.
arXiv Detail & Related papers (2024-06-07T09:48:11Z) - Revisiting Demonstration Selection Strategies in In-Context Learning [66.11652803887284]
Large language models (LLMs) have shown an impressive ability to perform a wide range of tasks using in-context learning (ICL)
In this work, we first revisit the factors contributing to this variance from both data and model aspects, and find that the choice of demonstration is both data- and model-dependent.
We propose a data- and model-dependent demonstration selection method, textbfTopK + ConE, based on the assumption that textitthe performance of a demonstration positively correlates with its contribution to the model's understanding of the test samples.
arXiv Detail & Related papers (2024-01-22T16:25:27Z) - Online Variational Sequential Monte Carlo [49.97673761305336]
We build upon the variational sequential Monte Carlo (VSMC) method, which provides computationally efficient and accurate model parameter estimation and Bayesian latent-state inference.
Online VSMC is capable of performing efficiently, entirely on-the-fly, both parameter estimation and particle proposal adaptation.
arXiv Detail & Related papers (2023-12-19T21:45:38Z) - Evaluating Representations with Readout Model Switching [19.907607374144167]
In this paper, we propose to use the Minimum Description Length (MDL) principle to devise an evaluation metric.
We design a hybrid discrete and continuous-valued model space for the readout models and employ a switching strategy to combine their predictions.
The proposed metric can be efficiently computed with an online method and we present results for pre-trained vision encoders of various architectures.
arXiv Detail & Related papers (2023-02-19T14:08:01Z) - Efficient Propagation of Uncertainty via Reordering Monte Carlo Samples [0.7087237546722617]
Uncertainty propagation is a technique to determine model output uncertainties based on the uncertainty in its input variables.
In this work, we investigate the hypothesis that while all samples are useful on average, some samples must be more useful than others.
We introduce a methodology to adaptively reorder MC samples and show how it results in reduction of computational expense of UP processes.
arXiv Detail & Related papers (2023-02-09T21:28:15Z) - Low-variance estimation in the Plackett-Luce model via quasi-Monte Carlo
sampling [58.14878401145309]
We develop a novel approach to producing more sample-efficient estimators of expectations in the PL model.
We illustrate our findings both theoretically and empirically using real-world recommendation data from Amazon Music and the Yahoo learning-to-rank challenge.
arXiv Detail & Related papers (2022-05-12T11:15:47Z) - PSD Representations for Effective Probability Models [117.35298398434628]
We show that a recently proposed class of positive semi-definite (PSD) models for non-negative functions is particularly suited to this end.
We characterize both approximation and generalization capabilities of PSD models, showing that they enjoy strong theoretical guarantees.
Our results open the way to applications of PSD models to density estimation, decision theory and inference.
arXiv Detail & Related papers (2021-06-30T15:13:39Z) - Community Detection in the Stochastic Block Model by Mixed Integer
Programming [3.8073142980733]
Degree-Corrected Block Model (DCSBM) is a popular model to generate random graphs with community structure given an expected degree sequence.
Standard approach of community detection based on the DCSBM is to search for the model parameters that are the most likely to have produced the observed network data through maximum likelihood estimation (MLE)
We present mathematical programming formulations and exact solution methods that can provably find the model parameters and community assignments of maximum likelihood given an observed graph.
arXiv Detail & Related papers (2021-01-26T22:04:40Z) - Evaluating the Disentanglement of Deep Generative Models through
Manifold Topology [66.06153115971732]
We present a method for quantifying disentanglement that only uses the generative model.
We empirically evaluate several state-of-the-art models across multiple datasets.
arXiv Detail & Related papers (2020-06-05T20:54:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.