Automated extraction of mutual independence patterns using Bayesian
comparison of partition models
- URL: http://arxiv.org/abs/2001.05407v1
- Date: Wed, 15 Jan 2020 16:21:48 GMT
- Title: Automated extraction of mutual independence patterns using Bayesian
comparison of partition models
- Authors: Guillaume Marrelec and Alain Giron
- Abstract summary: Mutual independence is a key concept in statistics that characterizes the structural relationships between variables.
Existing methods to investigate mutual independence rely on the definition of two competing models.
We propose a general Markov chain Monte Carlo (MCMC) algorithm to numerically approximate the posterior distribution on the space of all patterns of mutual independence.
- Score: 7.6146285961466
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Mutual independence is a key concept in statistics that characterizes the
structural relationships between variables. Existing methods to investigate
mutual independence rely on the definition of two competing models, one being
nested into the other and used to generate a null distribution for a statistic
of interest, usually under the asymptotic assumption of large sample size. As
such, these methods have a very restricted scope of application. In the present
manuscript, we propose to change the investigation of mutual independence from
a hypothesis-driven task that can only be applied in very specific cases to a
blind and automated search within patterns of mutual independence. To this end,
we treat the issue as one of model comparison that we solve in a Bayesian
framework. We show the relationship between such an approach and existing
methods in the case of multivariate normal distributions as well as
cross-classified multinomial distributions. We propose a general Markov chain
Monte Carlo (MCMC) algorithm to numerically approximate the posterior
distribution on the space of all patterns of mutual independence. The relevance
of the method is demonstrated on synthetic data as well as two real datasets,
showing the unique insight provided by this approach.
Related papers
- Theory on Score-Mismatched Diffusion Models and Zero-Shot Conditional Samplers [49.97755400231656]
We present the first performance guarantee with explicit dimensional general score-mismatched diffusion samplers.
We show that score mismatches result in an distributional bias between the target and sampling distributions, proportional to the accumulated mismatch between the target and training distributions.
This result can be directly applied to zero-shot conditional samplers for any conditional model, irrespective of measurement noise.
arXiv Detail & Related papers (2024-10-17T16:42:12Z) - Convergence of Score-Based Discrete Diffusion Models: A Discrete-Time Analysis [56.442307356162864]
We study the theoretical aspects of score-based discrete diffusion models under the Continuous Time Markov Chain (CTMC) framework.
We introduce a discrete-time sampling algorithm in the general state space $[S]d$ that utilizes score estimators at predefined time points.
Our convergence analysis employs a Girsanov-based method and establishes key properties of the discrete score function.
arXiv Detail & Related papers (2024-10-03T09:07:13Z) - Scalable Counterfactual Distribution Estimation in Multivariate Causal
Models [12.88471300865496]
We consider the problem of estimating the counterfactual joint distribution of multiple quantities of interests in a multivariate causal model.
We propose a method that alleviates both issues simultaneously by leveraging a robust latent one-dimensional subspace.
We demonstrate the advantages of our approach over existing methods on both synthetic and real-world data.
arXiv Detail & Related papers (2023-11-02T01:45:44Z) - MINDE: Mutual Information Neural Diffusion Estimation [7.399561232927219]
We present a new method for the estimation of Mutual Information (MI) between random variables.
We use score-based diffusion models to estimate the Kullback Leibler divergence between two densities as a difference between their score functions.
As a by-product, our method also enables the estimation of the entropy of random variables.
arXiv Detail & Related papers (2023-10-13T11:47:41Z) - Wasserstein multivariate auto-regressive models for modeling distributional time series [0.0]
We propose a new auto-regressive model for the statistical analysis of multivariate distributional time series.
Results on the existence, uniqueness and stationarity of the solution of such a model are provided.
To shed some light on the benefits of our approach for real data analysis, we also apply this methodology to a data set made of observations from age distribution in different countries.
arXiv Detail & Related papers (2022-07-12T10:18:36Z) - Nonparametric Conditional Local Independence Testing [69.31200003384122]
Conditional local independence is an independence relation among continuous time processes.
No nonparametric test of conditional local independence has been available.
We propose such a nonparametric test based on double machine learning.
arXiv Detail & Related papers (2022-03-25T10:31:02Z) - Sampling from Arbitrary Functions via PSD Models [55.41644538483948]
We take a two-step approach by first modeling the probability distribution and then sampling from that model.
We show that these models can approximate a large class of densities concisely using few evaluations, and present a simple algorithm to effectively sample from these models.
arXiv Detail & Related papers (2021-10-20T12:25:22Z) - Accounting for Unobserved Confounding in Domain Generalization [107.0464488046289]
This paper investigates the problem of learning robust, generalizable prediction models from a combination of datasets.
Part of the challenge of learning robust models lies in the influence of unobserved confounders.
We demonstrate the empirical performance of our approach on healthcare data from different modalities.
arXiv Detail & Related papers (2020-07-21T08:18:06Z) - Learning Ising models from one or multiple samples [26.00403702328348]
We provide guarantees for one-sample estimation, quantifying the estimation error in terms of the metric entropy of a family of interaction matrices.
Our technical approach benefits from sparsifying a model's interaction network, conditioning on subsets of variables that make the dependencies in the resulting conditional distribution sufficiently weak.
arXiv Detail & Related papers (2020-04-20T15:17:05Z) - GANs with Conditional Independence Graphs: On Subadditivity of
Probability Divergences [70.30467057209405]
Generative Adversarial Networks (GANs) are modern methods to learn the underlying distribution of a data set.
GANs are designed in a model-free fashion where no additional information about the underlying distribution is available.
We propose a principled design of a model-based GAN that uses a set of simple discriminators on the neighborhoods of the Bayes-net/MRF.
arXiv Detail & Related papers (2020-03-02T04:31:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.