Latent Space Model for Higher-order Networks and Generalized Tensor
Decomposition
- URL: http://arxiv.org/abs/2106.16042v1
- Date: Wed, 30 Jun 2021 13:11:17 GMT
- Title: Latent Space Model for Higher-order Networks and Generalized Tensor
Decomposition
- Authors: Zhongyuan Lyu and Dong Xia and Yuan Zhang
- Abstract summary: We introduce a unified framework, formulated as general latent space models, to study complex higher-order network interactions.
We formulate the relationship between the latent positions and the observed data via a generalized multilinear kernel as the link function.
We demonstrate the effectiveness of our method on synthetic data.
- Score: 18.07071669486882
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We introduce a unified framework, formulated as general latent space models,
to study complex higher-order network interactions among multiple entities. Our
framework covers several popular models in recent network analysis literature,
including mixture multi-layer latent space model and hypergraph latent space
model. We formulate the relationship between the latent positions and the
observed data via a generalized multilinear kernel as the link function. While
our model enjoys decent generality, its maximum likelihood parameter estimation
is also convenient via a generalized tensor decomposition procedure.We propose
a novel algorithm using projected gradient descent on Grassmannians. We also
develop original theoretical guarantees for our algorithm. First, we show its
linear convergence under mild conditions. Second, we establish finite-sample
statistical error rates of latent position estimation, determined by the signal
strength, degrees of freedom and the smoothness of link function, for both
general and specific latent space models. We demonstrate the effectiveness of
our method on synthetic data. We also showcase the merit of our method on two
real-world datasets that are conventionally described by different specific
models in producing meaningful and interpretable parameter estimations and
accurate link prediction. We demonstrate the effectiveness of our method on
synthetic data. We also showcase the merit of our method on two real-world
datasets that are conventionally described by different specific models in
producing meaningful and interpretable parameter estimations and accurate link
prediction.
Related papers
- Sub-graph Based Diffusion Model for Link Prediction [43.15741675617231]
Denoising Diffusion Probabilistic Models (DDPMs) represent a contemporary class of generative models with exceptional qualities.
We build a novel generative model for link prediction using a dedicated design to decompose the likelihood estimation process via the Bayesian formula.
Our proposed method presents numerous advantages: (1) transferability across datasets without retraining, (2) promising generalization on limited training data, and (3) robustness against graph adversarial attacks.
arXiv Detail & Related papers (2024-09-13T02:23:55Z) - Latent diffusion models for parameterization and data assimilation of facies-based geomodels [0.0]
Diffusion models are trained to generate new geological realizations from input fields characterized by random noise.
Latent diffusion models are shown to provide realizations that are visually consistent with samples from geomodeling software.
arXiv Detail & Related papers (2024-06-21T01:32:03Z) - Diffusion posterior sampling for simulation-based inference in tall data settings [53.17563688225137]
Simulation-based inference ( SBI) is capable of approximating the posterior distribution that relates input parameters to a given observation.
In this work, we consider a tall data extension in which multiple observations are available to better infer the parameters of the model.
We compare our method to recently proposed competing approaches on various numerical experiments and demonstrate its superiority in terms of numerical stability and computational cost.
arXiv Detail & Related papers (2024-04-11T09:23:36Z) - Latent Semantic Consensus For Deterministic Geometric Model Fitting [109.44565542031384]
We propose an effective method called Latent Semantic Consensus (LSC)
LSC formulates the model fitting problem into two latent semantic spaces based on data points and model hypotheses.
LSC is able to provide consistent and reliable solutions within only a few milliseconds for general multi-structural model fitting.
arXiv Detail & Related papers (2024-03-11T05:35:38Z) - Towards Theoretical Understandings of Self-Consuming Generative Models [56.84592466204185]
This paper tackles the emerging challenge of training generative models within a self-consuming loop.
We construct a theoretical framework to rigorously evaluate how this training procedure impacts the data distributions learned by future models.
We present results for kernel density estimation, delivering nuanced insights such as the impact of mixed data training on error propagation.
arXiv Detail & Related papers (2024-02-19T02:08:09Z) - Joint Distributional Learning via Cramer-Wold Distance [0.7614628596146602]
We introduce the Cramer-Wold distance regularization, which can be computed in a closed-form, to facilitate joint distributional learning for high-dimensional datasets.
We also introduce a two-step learning method to enable flexible prior modeling and improve the alignment between the aggregated posterior and the prior distribution.
arXiv Detail & Related papers (2023-10-25T05:24:23Z) - VTAE: Variational Transformer Autoencoder with Manifolds Learning [144.0546653941249]
Deep generative models have demonstrated successful applications in learning non-linear data distributions through a number of latent variables.
The nonlinearity of the generator implies that the latent space shows an unsatisfactory projection of the data space, which results in poor representation learning.
We show that geodesics and accurate computation can substantially improve the performance of deep generative models.
arXiv Detail & Related papers (2023-04-03T13:13:19Z) - Spatially-Varying Bayesian Predictive Synthesis for Flexible and
Interpretable Spatial Prediction [6.07227513262407]
We propose a novel methodology that captures spatially-varying model uncertainty, which we call spatial Bayesian predictive synthesis.
We show that our proposed spatial Bayesian predictive synthesis outperforms standard spatial models and advanced machine learning methods.
arXiv Detail & Related papers (2022-03-10T07:16:29Z) - Post-mortem on a deep learning contest: a Simpson's paradox and the
complementary roles of scale metrics versus shape metrics [61.49826776409194]
We analyze a corpus of models made publicly-available for a contest to predict the generalization accuracy of neural network (NN) models.
We identify what amounts to a Simpson's paradox: where "scale" metrics perform well overall but perform poorly on sub partitions of the data.
We present two novel shape metrics, one data-independent, and the other data-dependent, which can predict trends in the test accuracy of a series of NNs.
arXiv Detail & Related papers (2021-06-01T19:19:49Z) - GELATO: Geometrically Enriched Latent Model for Offline Reinforcement
Learning [54.291331971813364]
offline reinforcement learning approaches can be divided into proximal and uncertainty-aware methods.
In this work, we demonstrate the benefit of combining the two in a latent variational model.
Our proposed metrics measure both the quality of out of distribution samples as well as the discrepancy of examples in the data.
arXiv Detail & Related papers (2021-02-22T19:42:40Z) - Hierarchical regularization networks for sparsification based learning
on noisy datasets [0.0]
hierarchy follows from approximation spaces identified at successively finer scales.
For promoting model generalization at each scale, we also introduce a novel, projection based penalty operator across multiple dimension.
Results show the performance of the approach as a data reduction and modeling strategy on both synthetic and real datasets.
arXiv Detail & Related papers (2020-06-09T18:32:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.