Multi-constrained Symmetric Nonnegative Latent Factor Analysis for
Accurately Representing Large-scale Undirected Weighted Networks
- URL: http://arxiv.org/abs/2306.03911v1
- Date: Tue, 6 Jun 2023 14:13:16 GMT
- Title: Multi-constrained Symmetric Nonnegative Latent Factor Analysis for
Accurately Representing Large-scale Undirected Weighted Networks
- Authors: Yurong Zhong, Zhe Xie, Weiling Li, and Xin Luo
- Abstract summary: An Undirected Weighted Network (UWN) is frequently encountered in a big-data-related application.
An analysis model should carefully consider its symmetric-topology for describing an UWN's intrinsic symmetry.
This paper proposes a Multi-constrained Symmetric Nonnegative Latent-factor-analysis model with two-fold ideas.
- Score: 2.1797442801107056
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: An Undirected Weighted Network (UWN) is frequently encountered in a
big-data-related application concerning the complex interactions among numerous
nodes, e.g., a protein interaction network from a bioinformatics application. A
Symmetric High-Dimensional and Incomplete (SHDI) matrix can smoothly illustrate
such an UWN, which contains rich knowledge like node interaction behaviors and
local complexes. To extract desired knowledge from an SHDI matrix, an analysis
model should carefully consider its symmetric-topology for describing an UWN's
intrinsic symmetry. Representation learning to an UWN borrows the success of a
pyramid of symmetry-aware models like a Symmetric Nonnegative Matrix
Factorization (SNMF) model whose objective function utilizes a sole Latent
Factor (LF) matrix for representing SHDI's symmetry rigorously. However, they
suffer from the following drawbacks: 1) their computational complexity is high;
and 2) their modeling strategy narrows their representation features, making
them suffer from low learning ability. Aiming at addressing above critical
issues, this paper proposes a Multi-constrained Symmetric Nonnegative
Latent-factor-analysis (MSNL) model with two-fold ideas: 1) introducing
multi-constraints composed of multiple LF matrices, i.e., inequality and
equality ones into a data-density-oriented objective function for precisely
representing the intrinsic symmetry of an SHDI matrix with broadened feature
space; and 2) implementing an Alternating Direction Method of Multipliers
(ADMM)-incorporated learning scheme for precisely solving such a
multi-constrained model. Empirical studies on three SHDI matrices from a real
bioinformatics or industrial application demonstrate that the proposed MSNL
model achieves stronger representation learning ability to an SHDI matrix than
state-of-the-art models do.
Related papers
- Sample Complexity Characterization for Linear Contextual MDPs [67.79455646673762]
Contextual decision processes (CMDPs) describe a class of reinforcement learning problems in which the transition kernels and reward functions can change over time with different MDPs indexed by a context variable.
CMDPs serve as an important framework to model many real-world applications with time-varying environments.
We study CMDPs under two linear function approximation models: Model I with context-varying representations and common linear weights for all contexts; and Model II with common representations for all contexts and context-varying linear weights.
arXiv Detail & Related papers (2024-02-05T03:25:04Z) - Large-scale gradient-based training of Mixtures of Factor Analyzers [67.21722742907981]
This article contributes both a theoretical analysis as well as a new method for efficient high-dimensional training by gradient descent.
We prove that MFA training and inference/sampling can be performed based on precision matrices, which does not require matrix inversions after training is completed.
Besides the theoretical analysis and matrices, we apply MFA to typical image datasets such as SVHN and MNIST, and demonstrate the ability to perform sample generation and outlier detection.
arXiv Detail & Related papers (2023-08-26T06:12:33Z) - Proximal Symmetric Non-negative Latent Factor Analysis: A Novel Approach
to Highly-Accurate Representation of Undirected Weighted Networks [2.1797442801107056]
Undirected Weighted Network (UWN) is commonly found in big data-related applications.
Existing models fail in either modeling its intrinsic symmetry or low-data density.
Proximal Symmetric Nonnegative Latent-factor-analysis model is proposed.
arXiv Detail & Related papers (2023-06-06T13:03:24Z) - FAENet: Frame Averaging Equivariant GNN for Materials Modeling [123.19473575281357]
We introduce a flexible framework relying on frameaveraging (SFA) to make any model E(3)-equivariant or invariant through data transformations.
We prove the validity of our method theoretically and empirically demonstrate its superior accuracy and computational scalability in materials modeling.
arXiv Detail & Related papers (2023-04-28T21:48:31Z) - A Constraints Fusion-induced Symmetric Nonnegative Matrix Factorization
Approach for Community Detection [6.573829734173933]
Community is a fundamental and critical characteristic of an undirected social network.
This paper proposes a novel Constraints Fusion-induced Symmetric Nonnegative Matrix Factorization model.
It significantly outperforms state-of-the-art models in achieving highly-accurate community detection results.
arXiv Detail & Related papers (2023-02-23T15:52:14Z) - Learning Graphical Factor Models with Riemannian Optimization [70.13748170371889]
This paper proposes a flexible algorithmic framework for graph learning under low-rank structural constraints.
The problem is expressed as penalized maximum likelihood estimation of an elliptical distribution.
We leverage geometries of positive definite matrices and positive semi-definite matrices of fixed rank that are well suited to elliptical models.
arXiv Detail & Related papers (2022-10-21T13:19:45Z) - An Unconstrained Symmetric Nonnegative Latent Factor Analysis for
Large-scale Undirected Weighted Networks [0.22940141855172036]
Large-scale undirected weighted networks are usually found in big data-related research fields.
A symmetric non-negative latent-factor-analysis model is able to efficiently extract latent factors from an SHDI matrix.
This paper proposes an unconstrained symmetric nonnegative latent-factor-analysis model.
arXiv Detail & Related papers (2022-08-09T14:40:12Z) - A Multi-Metric Latent Factor Model for Analyzing High-Dimensional and
Sparse data [11.800988109180285]
High-dimensional and sparse (HiDS) matrices are omnipresent in a variety of big data-related applications.
Current LFA-based models mainly focus on a single-metric representation, where the representation strategy designed for the approximation Loss function is fixed and exclusive.
We in this paper propose a multi-metric latent factor (MMLF) model.
Our proposed MMLF enjoys the merits originated from a set of disparate metric spaces all at once, achieving the comprehensive and unbiased representation of HiDS matrices.
arXiv Detail & Related papers (2022-04-16T15:08:00Z) - Feature Weighted Non-negative Matrix Factorization [92.45013716097753]
We propose the Feature weighted Non-negative Matrix Factorization (FNMF) in this paper.
FNMF learns the weights of features adaptively according to their importances.
It can be solved efficiently with the suggested optimization algorithm.
arXiv Detail & Related papers (2021-03-24T21:17:17Z) - Information-theoretic limits of a multiview low-rank symmetric spiked
matrix model [19.738567726658875]
We consider a generalization of an important class of high-dimensional inference problems, namely spiked symmetric matrix models.
We rigorously establish the information-theoretic limits through the proof of single-letter formulas.
We improve the recently introduced adaptive method, so that it can be used to study low-rank models.
arXiv Detail & Related papers (2020-05-16T15:31:07Z) - Semiparametric Nonlinear Bipartite Graph Representation Learning with
Provable Guarantees [106.91654068632882]
We consider the bipartite graph and formalize its representation learning problem as a statistical estimation problem of parameters in a semiparametric exponential family distribution.
We show that the proposed objective is strongly convex in a neighborhood around the ground truth, so that a gradient descent-based method achieves linear convergence rate.
Our estimator is robust to any model misspecification within the exponential family, which is validated in extensive experiments.
arXiv Detail & Related papers (2020-03-02T16:40:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.