Asymptotic Theory of Eigenvectors for Latent Embeddings with Generalized Laplacian Matrices
- URL: http://arxiv.org/abs/2503.00640v1
- Date: Sat, 01 Mar 2025 22:22:42 GMT
- Title: Asymptotic Theory of Eigenvectors for Latent Embeddings with Generalized Laplacian Matrices
- Authors: Jianqing Fan, Yingying Fan, Jinchi Lv, Fan Yang, Diwen Yu,
- Abstract summary: dependency is a major bottleneck of new random matrix theory developments.<n>We suggest a new framework of eigenvectors for latent embeddings with generalized Laplacian matrices (ATE-GL)<n>We discuss some applications of the suggested ATE-GL framework and showcase its validity through some numerical examples.
- Score: 8.874743539416825
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Laplacian matrices are commonly employed in many real applications, encoding the underlying latent structural information such as graphs and manifolds. The use of the normalization terms naturally gives rise to random matrices with dependency. It is well-known that dependency is a major bottleneck of new random matrix theory (RMT) developments. To this end, in this paper, we formally introduce a class of generalized (and regularized) Laplacian matrices, which contains the Laplacian matrix and the random adjacency matrix as a specific case, and suggest the new framework of the asymptotic theory of eigenvectors for latent embeddings with generalized Laplacian matrices (ATE-GL). Our new theory is empowered by the tool of generalized quadratic vector equation for dealing with RMT under dependency, and delicate high-order asymptotic expansions of the empirical spiked eigenvectors and eigenvalues based on local laws. The asymptotic normalities established for both spiked eigenvectors and eigenvalues will enable us to conduct precise inference and uncertainty quantification for applications involving the generalized Laplacian matrices with flexibility. We discuss some applications of the suggested ATE-GL framework and showcase its validity through some numerical examples.
Related papers
- Matrix Ordering through Spectral and Nilpotent Structures in Totally Ordered Complex Number Fields [2.2533084621250143]
We develop a total ordering relation for complex numbers, enabling comparisons of the spectral components of general matrices with complex eigenvalues.<n>We establish a theoretical framework for majorization ordering with complex-valued functions.<n>We characterize Jordan blocks of matrix functions using a generalized dominance order for nilpotent components.
arXiv Detail & Related papers (2025-01-17T23:34:17Z) - Understanding Matrix Function Normalizations in Covariance Pooling through the Lens of Riemannian Geometry [63.694184882697435]
Global Covariance Pooling (GCP) has been demonstrated to improve the performance of Deep Neural Networks (DNNs) by exploiting second-order statistics of high-level representations.<n>This paper provides a comprehensive and unified understanding of the matrix logarithm and power from a Riemannian geometry perspective.
arXiv Detail & Related papers (2024-07-15T07:11:44Z) - Entrywise error bounds for low-rank approximations of kernel matrices [55.524284152242096]
We derive entrywise error bounds for low-rank approximations of kernel matrices obtained using the truncated eigen-decomposition.
A key technical innovation is a delocalisation result for the eigenvectors of the kernel matrix corresponding to small eigenvalues.
We validate our theory with an empirical study of a collection of synthetic and real-world datasets.
arXiv Detail & Related papers (2024-05-23T12:26:25Z) - Generalized unistochastic matrices [0.4604003661048266]
We measure a class of bistochastic matrices generalizing unistochastic matrices.
We show that the generalized unistochastic matrices is the whole Birkhoff polytope.
arXiv Detail & Related papers (2023-10-05T10:21:54Z) - The Inductive Bias of Flatness Regularization for Deep Matrix
Factorization [58.851514333119255]
This work takes the first step toward understanding the inductive bias of the minimum trace of the Hessian solutions in deep linear networks.
We show that for all depth greater than one, with the standard Isometry Property (RIP) on the measurements, minimizing the trace of Hessian is approximately equivalent to minimizing the Schatten 1-norm of the corresponding end-to-end matrix parameters.
arXiv Detail & Related papers (2023-06-22T23:14:57Z) - Trimmed Sampling Algorithm for the Noisy Generalized Eigenvalue Problem [0.0]
Solving the generalized eigenvalue problem is a useful method for finding energy eigenstates of large quantum systems.
The problem is especially bad when matrix elements are evaluated using methods and have significant error bars.
We introduce the trimmed sampling algorithm in order to solve this problem.
arXiv Detail & Related papers (2022-09-05T18:10:12Z) - Riemannian statistics meets random matrix theory: towards learning from
high-dimensional covariance matrices [2.352645870795664]
This paper shows that there seems to exist no practical method of computing the normalising factors associated with Riemannian Gaussian distributions on spaces of high-dimensional covariance matrices.
It is shown that this missing method comes from an unexpected new connection with random matrix theory.
Numerical experiments are conducted which demonstrate how this new approximation can unlock the difficulties which have impeded applications to real-world datasets.
arXiv Detail & Related papers (2022-03-01T03:16:50Z) - When Random Tensors meet Random Matrices [50.568841545067144]
This paper studies asymmetric order-$d$ spiked tensor models with Gaussian noise.
We show that the analysis of the considered model boils down to the analysis of an equivalent spiked symmetric textitblock-wise random matrix.
arXiv Detail & Related papers (2021-12-23T04:05:01Z) - Near optimal sample complexity for matrix and tensor normal models via
geodesic convexity [5.191641077435773]
We show nonasymptotic bounds for the error achieved by the maximum likelihood estimator (MLE) in several natural metrics.
In the same regimes as our sample complexity bounds, we show that an iterative procedure to compute the MLE known as the flip-flop algorithm converges linearly with high probability.
arXiv Detail & Related papers (2021-10-14T17:47:00Z) - Robust 1-bit Compressive Sensing with Partial Gaussian Circulant
Matrices and Generative Priors [54.936314353063494]
We provide recovery guarantees for a correlation-based optimization algorithm for robust 1-bit compressive sensing.
We make use of a practical iterative algorithm, and perform numerical experiments on image datasets to corroborate our results.
arXiv Detail & Related papers (2021-08-08T05:28:06Z) - Relations Between Adjacency and Modularity Graph Partitioning [0.3916094706589679]
This paper develops the exact linear relationship between the leading eigenvector of the unnormalized modularity matrix and the eigenvectors of the adjacency matrix.
There is also a complete proof of the equivalence between normalized adjacency clustering and normalized modularity clustering.
arXiv Detail & Related papers (2015-05-09T23:13:54Z) - Relative Error Bound Analysis for Nuclear Norm Regularized Matrix Completion [101.83262280224729]
We develop a relative error bound for nuclear norm regularized matrix completion.
We derive a relative upper bound for recovering the best low-rank approximation of the unknown matrix.
arXiv Detail & Related papers (2015-04-26T13:12:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.