Semiparametric Nonlinear Bipartite Graph Representation Learning with
Provable Guarantees
- URL: http://arxiv.org/abs/2003.01013v1
- Date: Mon, 2 Mar 2020 16:40:36 GMT
- Title: Semiparametric Nonlinear Bipartite Graph Representation Learning with
Provable Guarantees
- Authors: Sen Na, Yuwei Luo, Zhuoran Yang, Zhaoran Wang, Mladen Kolar
- Abstract summary: We consider the bipartite graph and formalize its representation learning problem as a statistical estimation problem of parameters in a semiparametric exponential family distribution.
We show that the proposed objective is strongly convex in a neighborhood around the ground truth, so that a gradient descent-based method achieves linear convergence rate.
Our estimator is robust to any model misspecification within the exponential family, which is validated in extensive experiments.
- Score: 106.91654068632882
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Graph representation learning is a ubiquitous task in machine learning where
the goal is to embed each vertex into a low-dimensional vector space. We
consider the bipartite graph and formalize its representation learning problem
as a statistical estimation problem of parameters in a semiparametric
exponential family distribution. The bipartite graph is assumed to be generated
by a semiparametric exponential family distribution, whose parametric component
is given by the proximity of outputs of two one-layer neural networks, while
nonparametric (nuisance) component is the base measure. Neural networks take
high-dimensional features as inputs and output embedding vectors. In this
setting, the representation learning problem is equivalent to recovering the
weight matrices. The main challenges of estimation arise from the nonlinearity
of activation functions and the nonparametric nuisance component of the
distribution. To overcome these challenges, we propose a pseudo-likelihood
objective based on the rank-order decomposition technique and focus on its
local geometry. We show that the proposed objective is strongly convex in a
neighborhood around the ground truth, so that a gradient descent-based method
achieves linear convergence rate. Moreover, we prove that the sample complexity
of the problem is linear in dimensions (up to logarithmic factors), which is
consistent with parametric Gaussian models. However, our estimator is robust to
any model misspecification within the exponential family, which is validated in
extensive experiments.
Related papers
- Learning Linear Causal Representations from Interventions under General
Nonlinear Mixing [52.66151568785088]
We prove strong identifiability results given unknown single-node interventions without access to the intervention targets.
This is the first instance of causal identifiability from non-paired interventions for deep neural network embeddings.
arXiv Detail & Related papers (2023-06-04T02:32:12Z) - Implicit Balancing and Regularization: Generalization and Convergence
Guarantees for Overparameterized Asymmetric Matrix Sensing [28.77440901439686]
A series of recent papers have begun to generalize this role for non-random Positive Semi-Defin (PSD) matrix sensing problems.
In this paper, we show that the trajectory of the gradient descent from small random measurements moves towards solutions that are both globally well.
arXiv Detail & Related papers (2023-03-24T19:05:52Z) - Learning Graphical Factor Models with Riemannian Optimization [70.13748170371889]
This paper proposes a flexible algorithmic framework for graph learning under low-rank structural constraints.
The problem is expressed as penalized maximum likelihood estimation of an elliptical distribution.
We leverage geometries of positive definite matrices and positive semi-definite matrices of fixed rank that are well suited to elliptical models.
arXiv Detail & Related papers (2022-10-21T13:19:45Z) - Low-Rank Covariance Completion for Graph Quilting with Applications to Functional Connectivity [8.500141848121782]
In many calcium imaging data sets, the full population of neurons is not recorded simultaneously, but instead in partially overlapping blocks.
In this paper, we introduce a novel two-step approach to Graph Quilting, which first imputes the nuclear structure matrix using low-rank co completion.
We show the efficacy of these methods for estimating connectivity from calcium imaging data.
arXiv Detail & Related papers (2022-09-17T08:03:46Z) - Error-Correcting Neural Networks for Two-Dimensional Curvature
Computation in the Level-Set Method [0.0]
We present an error-neural-modeling-based strategy for approximating two-dimensional curvature in the level-set method.
Our main contribution is a redesigned hybrid solver that relies on numerical schemes to enable machine-learning operations on demand.
arXiv Detail & Related papers (2022-01-22T05:14:40Z) - Minimax Estimation of Linear Functions of Eigenvectors in the Face of
Small Eigen-Gaps [95.62172085878132]
Eigenvector perturbation analysis plays a vital role in various statistical data science applications.
We develop a suite of statistical theory that characterizes the perturbation of arbitrary linear functions of an unknown eigenvector.
In order to mitigate a non-negligible bias issue inherent to the natural "plug-in" estimator, we develop de-biased estimators.
arXiv Detail & Related papers (2021-04-07T17:55:10Z) - Scaling Graph Clustering with Distributed Sketches [1.1011268090482575]
We present a method inspired by spectral clustering where we instead use matrix sketches derived from random dimension-reducing projections.
We show that our method produces embeddings that yield performant clustering results given a fully-dynamic block model stream.
We also discuss the effects of block model parameters upon the required dimensionality of the subsequent embeddings, and show how random projections could significantly improve the performance of graph clustering in distributed memory.
arXiv Detail & Related papers (2020-07-24T17:38:04Z) - Understanding Implicit Regularization in Over-Parameterized Single Index
Model [55.41685740015095]
We design regularization-free algorithms for the high-dimensional single index model.
We provide theoretical guarantees for the induced implicit regularization phenomenon.
arXiv Detail & Related papers (2020-07-16T13:27:47Z) - Multipole Graph Neural Operator for Parametric Partial Differential
Equations [57.90284928158383]
One of the main challenges in using deep learning-based methods for simulating physical systems is formulating physics-based data.
We propose a novel multi-level graph neural network framework that captures interaction at all ranges with only linear complexity.
Experiments confirm our multi-graph network learns discretization-invariant solution operators to PDEs and can be evaluated in linear time.
arXiv Detail & Related papers (2020-06-16T21:56:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.