Multi-task Learning for Gaussian Graphical Regressions with High
Dimensional Covariates
- URL: http://arxiv.org/abs/2205.10672v1
- Date: Sat, 21 May 2022 20:48:51 GMT
- Title: Multi-task Learning for Gaussian Graphical Regressions with High
Dimensional Covariates
- Authors: Jingfei Zhang and Yi Li
- Abstract summary: We propose a multi-task learning estimator for fitting Gaussian graphical regression models.
For computation, we consider an efficient augmented Lagrangian algorithm, which solves subproblems with a semi-smooth Newton method.
We show that the error rate of the multi-task learning based estimates has much improvement over that of the separate node-wise lasso estimates.
- Score: 5.726899123970559
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Gaussian graphical regression is a powerful means that regresses the
precision matrix of a Gaussian graphical model on covariates, permitting the
numbers of the response variables and covariates to far exceed the sample size.
Model fitting is typically carried out via separate node-wise lasso
regressions, ignoring the network-induced structure among these regressions.
Consequently, the error rate is high, especially when the number of nodes is
large. We propose a multi-task learning estimator for fitting Gaussian
graphical regression models; we design a cross-task group sparsity penalty and
a within task element-wise sparsity penalty, which govern the sparsity of
active covariates and their effects on the graph, respectively. For
computation, we consider an efficient augmented Lagrangian algorithm, which
solves subproblems with a semi-smooth Newton method. For theory, we show that
the error rate of the multi-task learning based estimates has much improvement
over that of the separate node-wise lasso estimates, because the cross-task
penalty borrows information across tasks. To address the main challenge that
the tasks are entangled in a complicated correlation structure, we establish a
new tail probability bound for correlated heavy-tailed (sub-exponential)
variables with an arbitrary correlation structure, a useful theoretical result
in its own right. Finally, the utility of our method is demonstrated through
simulations as well as an application to a gene co-expression network study
with brain cancer patients.
Related papers
- An Optimal Transport Approach for Network Regression [0.6238182916866519]
We build upon recent developments in generalized regression models on metric spaces based on Fr'echet means.
We propose a network regression method using the Wasserstein metric.
arXiv Detail & Related papers (2024-06-18T02:03:07Z) - Distributed High-Dimensional Quantile Regression: Estimation Efficiency and Support Recovery [0.0]
We focus on distributed estimation and support recovery for high-dimensional linear quantile regression.
We transform the original quantile regression into the least-squares optimization.
An efficient algorithm is developed, which enjoys high computation and communication efficiency.
arXiv Detail & Related papers (2024-05-13T08:32:22Z) - Neural Network-Based Score Estimation in Diffusion Models: Optimization
and Generalization [12.812942188697326]
Diffusion models have emerged as a powerful tool rivaling GANs in generating high-quality samples with improved fidelity, flexibility, and robustness.
A key component of these models is to learn the score function through score matching.
Despite empirical success on various tasks, it remains unclear whether gradient-based algorithms can learn the score function with a provable accuracy.
arXiv Detail & Related papers (2024-01-28T08:13:56Z) - Deep Generative Symbolic Regression [83.04219479605801]
Symbolic regression aims to discover concise closed-form mathematical equations from data.
Existing methods, ranging from search to reinforcement learning, fail to scale with the number of input variables.
We propose an instantiation of our framework, Deep Generative Symbolic Regression.
arXiv Detail & Related papers (2023-12-30T17:05:31Z) - Learning to Reweight for Graph Neural Network [63.978102332612906]
Graph Neural Networks (GNNs) show promising results for graph tasks.
Existing GNNs' generalization ability will degrade when there exist distribution shifts between testing and training graph data.
We propose a novel nonlinear graph decorrelation method, which can substantially improve the out-of-distribution generalization ability.
arXiv Detail & Related papers (2023-12-19T12:25:10Z) - Understanding Augmentation-based Self-Supervised Representation Learning
via RKHS Approximation and Regression [53.15502562048627]
Recent work has built the connection between self-supervised learning and the approximation of the top eigenspace of a graph Laplacian operator.
This work delves into a statistical analysis of augmentation-based pretraining.
arXiv Detail & Related papers (2023-06-01T15:18:55Z) - Multilayer Clustered Graph Learning [66.94201299553336]
We use contrastive loss as a data fidelity term, in order to properly aggregate the observed layers into a representative graph.
Experiments show that our method leads to a clustered clusters w.r.t.
We learn a clustering algorithm for solving clustering problems.
arXiv Detail & Related papers (2020-10-29T09:58:02Z) - Transfer Learning in Large-scale Gaussian Graphical Models with False
Discovery Rate Control [6.230751621285322]
An estimation algorithm, Trans-CLIME, is proposed and shown to attain a faster convergence rate than the minimax rate in the single study setting.
A significant decrease in prediction errors and a significant increase in power for link detection are observed.
arXiv Detail & Related papers (2020-10-21T14:39:14Z) - Multipole Graph Neural Operator for Parametric Partial Differential
Equations [57.90284928158383]
One of the main challenges in using deep learning-based methods for simulating physical systems is formulating physics-based data.
We propose a novel multi-level graph neural network framework that captures interaction at all ranges with only linear complexity.
Experiments confirm our multi-graph network learns discretization-invariant solution operators to PDEs and can be evaluated in linear time.
arXiv Detail & Related papers (2020-06-16T21:56:22Z) - Semiparametric Nonlinear Bipartite Graph Representation Learning with
Provable Guarantees [106.91654068632882]
We consider the bipartite graph and formalize its representation learning problem as a statistical estimation problem of parameters in a semiparametric exponential family distribution.
We show that the proposed objective is strongly convex in a neighborhood around the ground truth, so that a gradient descent-based method achieves linear convergence rate.
Our estimator is robust to any model misspecification within the exponential family, which is validated in extensive experiments.
arXiv Detail & Related papers (2020-03-02T16:40:36Z) - Residual Correlation in Graph Neural Network Regression [39.54530450932135]
We show that conditional independence assumption severely limits predictive power.
We address this problem with an interpretable and efficient framework.
Our framework achieves substantially higher accuracy than competing baselines.
arXiv Detail & Related papers (2020-02-19T16:32:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.