Learning Bayesian Networks through Birkhoff Polytope: A Relaxation
Method
- URL: http://arxiv.org/abs/2107.01658v1
- Date: Sun, 4 Jul 2021 15:04:02 GMT
- Title: Learning Bayesian Networks through Birkhoff Polytope: A Relaxation
Method
- Authors: Aramayis Dallakyan and Mohsen Pourahmadi
- Abstract summary: We establish a novel framework for learning a directed acyclic graph (DAG) when data are generated from a Gaussian, linear structural equation model.
For permutation matrix estimation, we propose a relaxation technique that avoids the NP-hard problem of order estimation.
Our framework recovers DAGs without the need for an expensive verification of the acyclicity constraint or enumeration of possible parent sets.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We establish a novel framework for learning a directed acyclic graph (DAG)
when data are generated from a Gaussian, linear structural equation model. It
consists of two parts: (1) introduce a permutation matrix as a new parameter
within a regularized Gaussian log-likelihood to represent variable ordering;
and (2) given the ordering, estimate the DAG structure through sparse Cholesky
factor of the inverse covariance matrix. For permutation matrix estimation, we
propose a relaxation technique that avoids the NP-hard combinatorial problem of
order estimation. Given an ordering, a sparse Cholesky factor is estimated
using a cyclic coordinatewise descent algorithm which decouples row-wise. Our
framework recovers DAGs without the need for an expensive verification of the
acyclicity constraint or enumeration of possible parent sets. We establish
numerical convergence of the algorithm, and consistency of the Cholesky factor
estimator when the order of variables is known. Through several simulated and
macro-economic datasets, we study the scope and performance of the proposed
methodology.
Related papers
- Riemannian Optimization for Non-convex Euclidean Distance Geometry with Global Recovery Guarantees [6.422262171968397]
Two algorithms are proposed to solve the Euclidean Distance Geometry problem.
First algorithm converges linearly to the true solution.
Second algorithm demonstrates strong numerical performance on both synthetic and real data.
arXiv Detail & Related papers (2024-10-08T21:19:22Z) - Induced Covariance for Causal Discovery in Linear Sparse Structures [55.2480439325792]
Causal models seek to unravel the cause-effect relationships among variables from observed data.
This paper introduces a novel causal discovery algorithm designed for settings in which variables exhibit linearly sparse relationships.
arXiv Detail & Related papers (2024-10-02T04:01:38Z) - An Efficient Algorithm for Clustered Multi-Task Compressive Sensing [60.70532293880842]
Clustered multi-task compressive sensing is a hierarchical model that solves multiple compressive sensing tasks.
The existing inference algorithm for this model is computationally expensive and does not scale well in high dimensions.
We propose a new algorithm that substantially accelerates model inference by avoiding the need to explicitly compute these covariance matrices.
arXiv Detail & Related papers (2023-09-30T15:57:14Z) - Recovering Simultaneously Structured Data via Non-Convex Iteratively
Reweighted Least Squares [0.8702432681310401]
We propose a new algorithm for recovering data that adheres to multiple, heterogeneous low-dimensional structures from linear observations.
We show that the IRLS method favorable in identifying low/comckuele state measurements.
arXiv Detail & Related papers (2023-06-08T06:35:47Z) - Classification of BCI-EEG based on augmented covariance matrix [0.0]
We propose a new framework based on the augmented covariance extracted from an autoregressive model to improve motor imagery classification.
We will test our approach on several datasets and several subjects using the MOABB framework.
arXiv Detail & Related papers (2023-02-09T09:04:25Z) - High-Dimensional Sparse Bayesian Learning without Covariance Matrices [66.60078365202867]
We introduce a new inference scheme that avoids explicit construction of the covariance matrix.
Our approach couples a little-known diagonal estimation result from numerical linear algebra with the conjugate gradient algorithm.
On several simulations, our method scales better than existing approaches in computation time and memory.
arXiv Detail & Related papers (2022-02-25T16:35:26Z) - Unfolding Projection-free SDP Relaxation of Binary Graph Classifier via
GDPA Linearization [59.87663954467815]
Algorithm unfolding creates an interpretable and parsimonious neural network architecture by implementing each iteration of a model-based algorithm as a neural layer.
In this paper, leveraging a recent linear algebraic theorem called Gershgorin disc perfect alignment (GDPA), we unroll a projection-free algorithm for semi-definite programming relaxation (SDR) of a binary graph.
Experimental results show that our unrolled network outperformed pure model-based graph classifiers, and achieved comparable performance to pure data-driven networks but using far fewer parameters.
arXiv Detail & Related papers (2021-09-10T07:01:15Z) - DAGs with No Curl: An Efficient DAG Structure Learning Approach [62.885572432958504]
Recently directed acyclic graph (DAG) structure learning is formulated as a constrained continuous optimization problem with continuous acyclicity constraints.
We propose a novel learning framework to model and learn the weighted adjacency matrices in the DAG space directly.
We show that our method provides comparable accuracy but better efficiency than baseline DAG structure learning methods on both linear and generalized structural equation models.
arXiv Detail & Related papers (2021-06-14T07:11:36Z) - EiGLasso for Scalable Sparse Kronecker-Sum Inverse Covariance Estimation [1.370633147306388]
We introduce EiGLasso, a highly scalable method for sparse Kronecker-sum inverse covariance estimation.
We show that EiGLasso achieves two to three orders-of-magnitude speed-up compared to the existing methods.
arXiv Detail & Related papers (2021-05-20T16:22:50Z) - Sparse Cholesky covariance parametrization for recovering latent
structure in ordered data [1.5349431582672617]
We focus on arbitrary zero patterns in the Cholesky factor of a covariance matrix.
For the ordered scenario, we propose a novel estimation method that is based on matrix loss penalization.
We give guidelines, based on the empirical results, about which of the methods analysed is more appropriate for each setting.
arXiv Detail & Related papers (2020-06-02T08:35:00Z) - Conjoined Dirichlet Process [63.89763375457853]
We develop a novel, non-parametric probabilistic biclustering method based on Dirichlet processes to identify biclusters with strong co-occurrence in both rows and columns.
We apply our method to two different applications, text mining and gene expression analysis, and demonstrate that our method improves bicluster extraction in many settings compared to existing approaches.
arXiv Detail & Related papers (2020-02-08T19:41:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.