DAGs with No Curl: An Efficient DAG Structure Learning Approach
- URL: http://arxiv.org/abs/2106.07197v1
- Date: Mon, 14 Jun 2021 07:11:36 GMT
- Title: DAGs with No Curl: An Efficient DAG Structure Learning Approach
- Authors: Yue Yu, Tian Gao, Naiyu Yin, Qiang Ji
- Abstract summary: Recently directed acyclic graph (DAG) structure learning is formulated as a constrained continuous optimization problem with continuous acyclicity constraints.
We propose a novel learning framework to model and learn the weighted adjacency matrices in the DAG space directly.
We show that our method provides comparable accuracy but better efficiency than baseline DAG structure learning methods on both linear and generalized structural equation models.
- Score: 62.885572432958504
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Recently directed acyclic graph (DAG) structure learning is formulated as a
constrained continuous optimization problem with continuous acyclicity
constraints and was solved iteratively through subproblem optimization. To
further improve efficiency, we propose a novel learning framework to model and
learn the weighted adjacency matrices in the DAG space directly. Specifically,
we first show that the set of weighted adjacency matrices of DAGs are
equivalent to the set of weighted gradients of graph potential functions, and
one may perform structure learning by searching in this equivalent set of DAGs.
To instantiate this idea, we propose a new algorithm, DAG-NoCurl, which solves
the optimization problem efficiently with a two-step procedure: 1) first we
find an initial cyclic solution to the optimization problem, and 2) then we
employ the Hodge decomposition of graphs and learn an acyclic graph by
projecting the cyclic graph to the gradient of a potential function.
Experimental studies on benchmark datasets demonstrate that our method provides
comparable accuracy but better efficiency than baseline DAG structure learning
methods on both linear and generalized structural equation models, often by
more than one order of magnitude.
Related papers
- $ψ$DAG: Projected Stochastic Approximation Iteration for DAG Structure Learning [6.612096312467342]
Learning the structure of Directed A Graphs (DAGs) presents a significant challenge due to the vast search space of possible graphs, which scales with the number of nodes.
Recent advancements have redefined this problem as a continuous optimization task by incorporating differentiable a exponentiallyity constraints.
We present a novel framework for learning DAGs, employing a Approximation approach integrated with Gradient Descent (SGD)-based optimization techniques.
arXiv Detail & Related papers (2024-10-31T12:13:11Z) - Non-negative Weighted DAG Structure Learning [12.139158398361868]
We address the problem of learning the true DAGs from nodal observations.
We propose a DAG recovery algorithm based on the method that is guaranteed to return ar.
arXiv Detail & Related papers (2024-09-12T09:41:29Z) - Recovering Linear Causal Models with Latent Variables via Cholesky
Factorization of Covariance Matrix [21.698480201955213]
We propose a DAG structure recovering algorithm, which is based on the Cholesky factorization of the covariance matrix of the observed data.
On synthetic and real-world datasets, the algorithm is significantly faster than previous methods and achieves the state-of-the-art performance.
arXiv Detail & Related papers (2023-11-01T17:27:49Z) - Linearization Algorithms for Fully Composite Optimization [61.20539085730636]
This paper studies first-order algorithms for solving fully composite optimization problems convex compact sets.
We leverage the structure of the objective by handling differentiable and non-differentiable separately, linearizing only the smooth parts.
arXiv Detail & Related papers (2023-02-24T18:41:48Z) - Multi-task Learning of Order-Consistent Causal Graphs [59.9575145128345]
We consider the problem of discovering $K related Gaussian acyclic graphs (DAGs)
Under multi-task learning setting, we propose a $l_1/l$-regularized maximum likelihood estimator (MLE) for learning $K$ linear structural equation models.
We theoretically show that the joint estimator, by leveraging data across related tasks, can achieve a better sample complexity for recovering the causal order.
arXiv Detail & Related papers (2021-11-03T22:10:18Z) - Learning Large DAGs by Combining Continuous Optimization and Feedback
Arc Set Heuristics [0.3553493344868413]
We propose two scalable NPs for learning DAGs in a linear structural equation case.
Our methods learn the DAG by alternating between unconstrained gradient descent-based step to optimize an objective function.
Thanks to this decoupling, our methods scale up beyond thousands of variables.
arXiv Detail & Related papers (2021-07-01T16:10:21Z) - A Bi-Level Framework for Learning to Solve Combinatorial Optimization on
Graphs [91.07247251502564]
We propose a hybrid approach to combine the best of the two worlds, in which a bi-level framework is developed with an upper-level learning method to optimize the graph.
Such a bi-level approach simplifies the learning on the original hard CO and can effectively mitigate the demand for model capacity.
arXiv Detail & Related papers (2021-06-09T09:18:18Z) - Leveraging Non-uniformity in First-order Non-convex Optimization [93.6817946818977]
Non-uniform refinement of objective functions leads to emphNon-uniform Smoothness (NS) and emphNon-uniform Lojasiewicz inequality (NL)
New definitions inspire new geometry-aware first-order methods that converge to global optimality faster than the classical $Omega (1/t2)$ lower bounds.
arXiv Detail & Related papers (2021-05-13T04:23:07Z) - Cogradient Descent for Bilinear Optimization [124.45816011848096]
We introduce a Cogradient Descent algorithm (CoGD) to address the bilinear problem.
We solve one variable by considering its coupling relationship with the other, leading to a synchronous gradient descent.
Our algorithm is applied to solve problems with one variable under the sparsity constraint.
arXiv Detail & Related papers (2020-06-16T13:41:54Z) - Learning DAGs without imposing acyclicity [0.6526824510982799]
We show that it is possible to learn a directed acyclic graph (DAG) from data without imposing the acyclicity constraint.
This approach is computationally efficient and is not affected by the explosion of complexity as in classical structural learning algorithms.
arXiv Detail & Related papers (2020-06-04T16:52:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.