Sparse Cholesky covariance parametrization for recovering latent
structure in ordered data
- URL: http://arxiv.org/abs/2006.01448v2
- Date: Wed, 19 Aug 2020 14:23:38 GMT
- Title: Sparse Cholesky covariance parametrization for recovering latent
structure in ordered data
- Authors: Irene C\'ordoba and Concha Bielza and Pedro Larra\~naga and Gherardo
Varando
- Abstract summary: We focus on arbitrary zero patterns in the Cholesky factor of a covariance matrix.
For the ordered scenario, we propose a novel estimation method that is based on matrix loss penalization.
We give guidelines, based on the empirical results, about which of the methods analysed is more appropriate for each setting.
- Score: 1.5349431582672617
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The sparse Cholesky parametrization of the inverse covariance matrix can be
interpreted as a Gaussian Bayesian network; however its counterpart, the
covariance Cholesky factor, has received, with few notable exceptions, little
attention so far, despite having a natural interpretation as a hidden variable
model for ordered signal data. To fill this gap, in this paper we focus on
arbitrary zero patterns in the Cholesky factor of a covariance matrix. We
discuss how these models can also be extended, in analogy with Gaussian
Bayesian networks, to data where no apparent order is available. For the
ordered scenario, we propose a novel estimation method that is based on matrix
loss penalization, as opposed to the existing regression-based approaches. The
performance of this sparse model for the Cholesky factor, together with our
novel estimator, is assessed in a simulation setting, as well as over spatial
and temporal real data where a natural ordering arises among the variables. We
give guidelines, based on the empirical results, about which of the methods
analysed is more appropriate for each setting.
Related papers
- Induced Covariance for Causal Discovery in Linear Sparse Structures [55.2480439325792]
Causal models seek to unravel the cause-effect relationships among variables from observed data.
This paper introduces a novel causal discovery algorithm designed for settings in which variables exhibit linearly sparse relationships.
arXiv Detail & Related papers (2024-10-02T04:01:38Z) - Intrinsic Bayesian Cramér-Rao Bound with an Application to Covariance Matrix Estimation [49.67011673289242]
This paper presents a new performance bound for estimation problems where the parameter to estimate lies in a smooth manifold.
It induces a geometry for the parameter manifold, as well as an intrinsic notion of the estimation error measure.
arXiv Detail & Related papers (2023-11-08T15:17:13Z) - A probabilistic, data-driven closure model for RANS simulations with aleatoric, model uncertainty [1.8416014644193066]
We propose a data-driven, closure model for Reynolds-averaged Navier-Stokes (RANS) simulations that incorporates aleatoric, model uncertainty.
A fully Bayesian formulation is proposed, combined with a sparsity-inducing prior in order to identify regions in the problem domain where the parametric closure is insufficient.
arXiv Detail & Related papers (2023-07-05T16:53:31Z) - Score-based Continuous-time Discrete Diffusion Models [102.65769839899315]
We extend diffusion models to discrete variables by introducing a Markov jump process where the reverse process denoises via a continuous-time Markov chain.
We show that an unbiased estimator can be obtained via simple matching the conditional marginal distributions.
We demonstrate the effectiveness of the proposed method on a set of synthetic and real-world music and image benchmarks.
arXiv Detail & Related papers (2022-11-30T05:33:29Z) - Equivariance Discovery by Learned Parameter-Sharing [153.41877129746223]
We study how to discover interpretable equivariances from data.
Specifically, we formulate this discovery process as an optimization problem over a model's parameter-sharing schemes.
Also, we theoretically analyze the method for Gaussian data and provide a bound on the mean squared gap between the studied discovery scheme and the oracle scheme.
arXiv Detail & Related papers (2022-04-07T17:59:19Z) - Learning Structured Gaussians to Approximate Deep Ensembles [10.055143995729415]
This paper proposes using a sparse-structured multivariate Gaussian to provide a closed-form approxorimator for dense image prediction tasks.
We capture the uncertainty and structured correlations in the predictions explicitly in a formal distribution, rather than implicitly through sampling alone.
We demonstrate the merits of our approach on monocular depth estimation and show that the advantages of our approach are obtained with comparable quantitative performance.
arXiv Detail & Related papers (2022-03-29T12:34:43Z) - Learning Bayesian Networks through Birkhoff Polytope: A Relaxation
Method [0.0]
We establish a novel framework for learning a directed acyclic graph (DAG) when data are generated from a Gaussian, linear structural equation model.
For permutation matrix estimation, we propose a relaxation technique that avoids the NP-hard problem of order estimation.
Our framework recovers DAGs without the need for an expensive verification of the acyclicity constraint or enumeration of possible parent sets.
arXiv Detail & Related papers (2021-07-04T15:04:02Z) - Model Fusion with Kullback--Leibler Divergence [58.20269014662046]
We propose a method to fuse posterior distributions learned from heterogeneous datasets.
Our algorithm relies on a mean field assumption for both the fused model and the individual dataset posteriors.
arXiv Detail & Related papers (2020-07-13T03:27:45Z) - Fitting Laplacian Regularized Stratified Gaussian Models [0.0]
We consider the problem of jointly estimating multiple related zero-mean Gaussian distributions from data.
We propose a distributed method that scales to large problems, and illustrate the efficacy of the method with examples in finance, radar signal processing, and weather forecasting.
arXiv Detail & Related papers (2020-05-04T18:00:59Z) - Asymptotic Analysis of an Ensemble of Randomly Projected Linear
Discriminants [94.46276668068327]
In [1], an ensemble of randomly projected linear discriminants is used to classify datasets.
We develop a consistent estimator of the misclassification probability as an alternative to the computationally-costly cross-validation estimator.
We also demonstrate the use of our estimator for tuning the projection dimension on both real and synthetic data.
arXiv Detail & Related papers (2020-04-17T12:47:04Z) - Generalized Gumbel-Softmax Gradient Estimator for Various Discrete
Random Variables [16.643346012854156]
Esting the gradients of nodes is one of the crucial research questions in the deep generative modeling community.
This paper proposes a general version of the Gumbel-Softmax estimator with continuous relaxation.
arXiv Detail & Related papers (2020-03-04T01:13:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.