Rigorous Explanation of Inference on Probabilistic Graphical Models
- URL: http://arxiv.org/abs/2004.10066v1
- Date: Tue, 21 Apr 2020 14:57:12 GMT
- Title: Rigorous Explanation of Inference on Probabilistic Graphical Models
- Authors: Yifei Liu, Chao Chen, Xi Zhang, Sihong Xie
- Abstract summary: We propose GraphShapley to integrate the decomposability of Shapley values, the structure of computation MRFs, and the iterative nature of BP inference.
On nine graphs, we demonstrate that GraphShapley provides sensible and practical explanations.
- Score: 17.96228289921288
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Probabilistic graphical models, such as Markov random fields (MRF), exploit
dependencies among random variables to model a rich family of joint probability
distributions. Sophisticated inference algorithms, such as belief propagation
(BP), can effectively compute the marginal posteriors. Nonetheless, it is still
difficult to interpret the inference outcomes for important human decision
making. There is no existing method to rigorously attribute the inference
outcomes to the contributing factors of the graphical models. Shapley values
provide an axiomatic framework, but naively computing or even approximating the
values on general graphical models is challenging and less studied. We propose
GraphShapley to integrate the decomposability of Shapley values, the structure
of MRFs, and the iterative nature of BP inference in a principled way for fast
Shapley value computation, that 1) systematically enumerates the important
contributions to the Shapley values of the explaining variables without
duplicate; 2) incrementally compute the contributions without starting from
scratches. We theoretically characterize GraphShapley regarding independence,
equal contribution, and additivity. On nine graphs, we demonstrate that
GraphShapley provides sensible and practical explanations.
Related papers
- Shapley Marginal Surplus for Strong Models [0.9831489366502301]
We show that while Shapley values might be accurate explainers of model predictions, machine learning models themselves are often poor explainers of the true data-generating process (DGP)
We introduce a novel variable importance algorithm, Shapley Marginal Surplus for Strong Models, that samples the space of possible models to come up with an inferential measure of feature importance.
arXiv Detail & Related papers (2024-08-16T17:06:07Z) - Goodness-of-Fit of Attributed Probabilistic Graph Generative Models [11.58149447373971]
We define goodness of fit in terms of the mean square contingency coefficient for random binary networks.
We apply these criteria to verify the representation capability of a probabilistic generative model for various popular types of graph models.
arXiv Detail & Related papers (2023-07-28T18:48:09Z) - Explaining the Uncertain: Stochastic Shapley Values for Gaussian Process
Models [15.715453687736028]
We present a novel approach for explaining Gaussian processes (GPs) that can utilize the full analytical covariance structure in GPs.
Our method is based on the popular solution concept of Shapley values extended to cooperative games, resulting in explanations that are random variables.
The GP explanations generated using our approach satisfy similar axioms to standard Shapley values and possess a tractable covariance function across features and data observations.
arXiv Detail & Related papers (2023-05-24T13:59:03Z) - Learning Graphical Factor Models with Riemannian Optimization [70.13748170371889]
This paper proposes a flexible algorithmic framework for graph learning under low-rank structural constraints.
The problem is expressed as penalized maximum likelihood estimation of an elliptical distribution.
We leverage geometries of positive definite matrices and positive semi-definite matrices of fixed rank that are well suited to elliptical models.
arXiv Detail & Related papers (2022-10-21T13:19:45Z) - Invariance Principle Meets Out-of-Distribution Generalization on Graphs [66.04137805277632]
Complex nature of graphs thwarts the adoption of the invariance principle for OOD generalization.
domain or environment partitions, which are often required by OOD methods, can be expensive to obtain for graphs.
We propose a novel framework to explicitly model this process using a contrastive strategy.
arXiv Detail & Related papers (2022-02-11T04:38:39Z) - Probabilistic Simplex Component Analysis [66.30587591100566]
PRISM is a probabilistic simplex component analysis approach to identifying the vertices of a data-circumscribing simplex from data.
The problem has a rich variety of applications, the most notable being hyperspectral unmixing in remote sensing and non-negative matrix factorization in machine learning.
arXiv Detail & Related papers (2021-03-18T05:39:00Z) - Explaining predictive models using Shapley values and non-parametric
vine copulas [2.6774008509840996]
We propose two new approaches for modelling the dependence between the features.
The performance of the proposed methods is evaluated on simulated data sets and a real data set.
Experiments demonstrate that the vine copula approaches give more accurate approximations to the true Shapley values than its competitors.
arXiv Detail & Related papers (2021-02-12T09:43:28Z) - Causal Shapley Values: Exploiting Causal Knowledge to Explain Individual
Predictions of Complex Models [6.423239719448169]
Shapley values are designed to attribute the difference between a model's prediction and an average baseline to the different features used as input to the model.
We show how these 'causal' Shapley values can be derived for general causal graphs without sacrificing any of their desirable properties.
arXiv Detail & Related papers (2020-11-03T11:11:36Z) - Probabilistic Circuits for Variational Inference in Discrete Graphical
Models [101.28528515775842]
Inference in discrete graphical models with variational methods is difficult.
Many sampling-based methods have been proposed for estimating Evidence Lower Bound (ELBO)
We propose a new approach that leverages the tractability of probabilistic circuit models, such as Sum Product Networks (SPN)
We show that selective-SPNs are suitable as an expressive variational distribution, and prove that when the log-density of the target model is aweighted the corresponding ELBO can be computed analytically.
arXiv Detail & Related papers (2020-10-22T05:04:38Z) - Multiplicative noise and heavy tails in stochastic optimization [62.993432503309485]
empirical optimization is central to modern machine learning, but its role in its success is still unclear.
We show that it commonly arises in parameters of discrete multiplicative noise due to variance.
A detailed analysis is conducted in which we describe on key factors, including recent step size, and data, all exhibit similar results on state-of-the-art neural network models.
arXiv Detail & Related papers (2020-06-11T09:58:01Z) - Block-Approximated Exponential Random Graphs [77.4792558024487]
An important challenge in the field of exponential random graphs (ERGs) is the fitting of non-trivial ERGs on large graphs.
We propose an approximative framework to such non-trivial ERGs that result in dyadic independence (i.e., edge independent) distributions.
Our methods are scalable to sparse graphs consisting of millions of nodes.
arXiv Detail & Related papers (2020-02-14T11:42:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.