Multi-task Learning of Order-Consistent Causal Graphs
- URL: http://arxiv.org/abs/2111.02545v1
- Date: Wed, 3 Nov 2021 22:10:18 GMT
- Title: Multi-task Learning of Order-Consistent Causal Graphs
- Authors: Xinshi Chen, Haoran Sun, Caleb Ellington, Eric Xing, Le Song
- Abstract summary: We consider the problem of discovering $K related Gaussian acyclic graphs (DAGs)
Under multi-task learning setting, we propose a $l_1/l$-regularized maximum likelihood estimator (MLE) for learning $K$ linear structural equation models.
We theoretically show that the joint estimator, by leveraging data across related tasks, can achieve a better sample complexity for recovering the causal order.
- Score: 59.9575145128345
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: We consider the problem of discovering $K$ related Gaussian directed acyclic
graphs (DAGs), where the involved graph structures share a consistent causal
order and sparse unions of supports. Under the multi-task learning setting, we
propose a $l_1/l_2$-regularized maximum likelihood estimator (MLE) for learning
$K$ linear structural equation models. We theoretically show that the joint
estimator, by leveraging data across related tasks, can achieve a better sample
complexity for recovering the causal order (or topological order) than separate
estimations. Moreover, the joint estimator is able to recover non-identifiable
DAGs, by estimating them together with some identifiable DAGs. Lastly, our
analysis also shows the consistency of union support recovery of the
structures. To allow practical implementation, we design a continuous
optimization problem whose optimizer is the same as the joint estimator and can
be approximated efficiently by an iterative algorithm. We validate the
theoretical analysis and the effectiveness of the joint estimator in
experiments.
Related papers
- $ψ$DAG: Projected Stochastic Approximation Iteration for DAG Structure Learning [6.612096312467342]
Learning the structure of Directed A Graphs (DAGs) presents a significant challenge due to the vast search space of possible graphs, which scales with the number of nodes.
Recent advancements have redefined this problem as a continuous optimization task by incorporating differentiable a exponentiallyity constraints.
We present a novel framework for learning DAGs, employing a Approximation approach integrated with Gradient Descent (SGD)-based optimization techniques.
arXiv Detail & Related papers (2024-10-31T12:13:11Z) - A Statistical Theory of Regularization-Based Continual Learning [10.899175512941053]
We provide a statistical analysis of regularization-based continual learning on a sequence of linear regression tasks.
We first derive the convergence rate for the oracle estimator obtained as if all data were available simultaneously.
A byproduct of our theoretical analysis is the equivalence between early stopping and generalized $ell$-regularization.
arXiv Detail & Related papers (2024-06-10T12:25:13Z) - Learning Directed Acyclic Graphs from Partial Orderings [9.387234607473054]
directed acyclic graphs (DAGs) are commonly used to model causal relationships among random variables.
In this paper, we consider the intermediate problem of learning DAGs when a partial causal ordering of variables is available.
We propose a general estimation framework for leveraging the partial ordering and present efficient estimation algorithms for low- and high-dimensional problems.
arXiv Detail & Related papers (2024-03-24T06:14:50Z) - Discovering Dynamic Causal Space for DAG Structure Learning [64.763763417533]
We propose a dynamic causal space for DAG structure learning, coined CASPER.
It integrates the graph structure into the score function as a new measure in the causal space to faithfully reflect the causal distance between estimated and ground truth DAG.
arXiv Detail & Related papers (2023-06-05T12:20:40Z) - Understanding Augmentation-based Self-Supervised Representation Learning
via RKHS Approximation and Regression [53.15502562048627]
Recent work has built the connection between self-supervised learning and the approximation of the top eigenspace of a graph Laplacian operator.
This work delves into a statistical analysis of augmentation-based pretraining.
arXiv Detail & Related papers (2023-06-01T15:18:55Z) - On Single-Objective Sub-Graph-Based Mutation for Solving the
Bi-Objective Minimum Spanning Tree Problem [0.0]
We contribute to the efficient approximation of the $mathcalNP$-hard multi-objective minimum spanning tree problem (moMST) adopting evolutionary computation.
We design several highly biased sub-graph-based mutation operators founded on the gained insights.
Our results confirm that the sub-graph based operators beat baseline algorithms from the literature.
arXiv Detail & Related papers (2023-05-31T22:35:17Z) - dotears: Scalable, consistent DAG estimation using observational and
interventional data [1.220743263007369]
Causal gene regulatory networks can be represented by directed acyclic graph (DAG)
We present $texttdotears$ [doo-tairs], a continuous optimization framework to infer a single causal structure.
We show that $texttdotears$ is a provably consistent estimator of the true DAG under mild assumptions.
arXiv Detail & Related papers (2023-05-30T17:03:39Z) - Amortized Inference for Causal Structure Learning [72.84105256353801]
Learning causal structure poses a search problem that typically involves evaluating structures using a score or independence test.
We train a variational inference model to predict the causal structure from observational/interventional data.
Our models exhibit robust generalization capabilities under substantial distribution shift.
arXiv Detail & Related papers (2022-05-25T17:37:08Z) - DAGs with No Curl: An Efficient DAG Structure Learning Approach [62.885572432958504]
Recently directed acyclic graph (DAG) structure learning is formulated as a constrained continuous optimization problem with continuous acyclicity constraints.
We propose a novel learning framework to model and learn the weighted adjacency matrices in the DAG space directly.
We show that our method provides comparable accuracy but better efficiency than baseline DAG structure learning methods on both linear and generalized structural equation models.
arXiv Detail & Related papers (2021-06-14T07:11:36Z) - Joint Network Topology Inference via Structured Fusion Regularization [70.30364652829164]
Joint network topology inference represents a canonical problem of learning multiple graph Laplacian matrices from heterogeneous graph signals.
We propose a general graph estimator based on a novel structured fusion regularization.
We show that the proposed graph estimator enjoys both high computational efficiency and rigorous theoretical guarantee.
arXiv Detail & Related papers (2021-03-05T04:42:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.