Quantitative deterministic equivalent of sample covariance matrices with
a general dependence structure
- URL: http://arxiv.org/abs/2211.13044v1
- Date: Wed, 23 Nov 2022 15:50:31 GMT
- Title: Quantitative deterministic equivalent of sample covariance matrices with
a general dependence structure
- Authors: Cl\'ement Chouard (UT3)
- Abstract summary: We prove quantitative bounds involving both the dimensions and the spectral parameter, in particular allowing it to get closer to the real positive semi-line.
As applications, we obtain a new bound for the convergence in Kolmogorov distance of the empirical spectral distributions of these general models.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study sample covariance matrices arising from rectangular random matrices
with i.i.d. columns. It was previously known that the resolvent of these
matrices admits a deterministic equivalent when the spectral parameter stays
bounded away from the real axis. We extend this work by proving quantitative
bounds involving both the dimensions and the spectral parameter, in particular
allowing it to get closer to the real positive semi-line. As applications, we
obtain a new bound for the convergence in Kolmogorov distance of the empirical
spectral distributions of these general models. We also apply our framework to
the problem of regularization of Random Features models in Machine Learning
without Gaussian hypothesis.
Related papers
- Entrywise error bounds for low-rank approximations of kernel matrices [55.524284152242096]
We derive entrywise error bounds for low-rank approximations of kernel matrices obtained using the truncated eigen-decomposition.
A key technical innovation is a delocalisation result for the eigenvectors of the kernel matrix corresponding to small eigenvalues.
We validate our theory with an empirical study of a collection of synthetic and real-world datasets.
arXiv Detail & Related papers (2024-05-23T12:26:25Z) - Convex Parameter Estimation of Perturbed Multivariate Generalized
Gaussian Distributions [18.95928707619676]
We propose a convex formulation with well-established properties for MGGD parameters.
The proposed framework is flexible as it combines a variety of regularizations for the precision matrix, the mean and perturbations.
Experiments show a more accurate precision and covariance matrix estimation with similar performance for the mean vector parameter.
arXiv Detail & Related papers (2023-12-12T18:08:04Z) - Intrinsic Bayesian Cramér-Rao Bound with an Application to Covariance Matrix Estimation [49.67011673289242]
This paper presents a new performance bound for estimation problems where the parameter to estimate lies in a smooth manifold.
It induces a geometry for the parameter manifold, as well as an intrinsic notion of the estimation error measure.
arXiv Detail & Related papers (2023-11-08T15:17:13Z) - Deterministic equivalent of the Conjugate Kernel matrix associated to
Artificial Neural Networks [0.0]
We show that the empirical spectral distribution of the Conjugate Kernel converges to a deterministic limit.
More precisely we obtain a deterministic equivalent for its Stieltjes transform and its resolvent, with quantitative bounds involving both the dimension and the spectral parameter.
arXiv Detail & Related papers (2023-06-09T12:31:59Z) - On confidence intervals for precision matrices and the
eigendecomposition of covariance matrices [20.20416580970697]
This paper tackles the challenge of computing confidence bounds on the individual entries of eigenvectors of a covariance matrix of fixed dimension.
We derive a method to bound the entries of the inverse covariance matrix, the so-called precision matrix.
As an application of these results, we demonstrate a new statistical test, which allows us to test for non-zero values of the precision matrix.
arXiv Detail & Related papers (2022-08-25T10:12:53Z) - An Equivalence Principle for the Spectrum of Random Inner-Product Kernel
Matrices with Polynomial Scalings [21.727073594338297]
This study is motivated by applications in machine learning and statistics.
We establish the weak limit of the empirical distribution of these random matrices in a scaling regime.
Our results can be characterized as the free additive convolution between a Marchenko-Pastur law and a semicircle law.
arXiv Detail & Related papers (2022-05-12T18:50:21Z) - When Random Tensors meet Random Matrices [50.568841545067144]
This paper studies asymmetric order-$d$ spiked tensor models with Gaussian noise.
We show that the analysis of the considered model boils down to the analysis of an equivalent spiked symmetric textitblock-wise random matrix.
arXiv Detail & Related papers (2021-12-23T04:05:01Z) - Spectral clustering under degree heterogeneity: a case for the random
walk Laplacian [83.79286663107845]
This paper shows that graph spectral embedding using the random walk Laplacian produces vector representations which are completely corrected for node degree.
In the special case of a degree-corrected block model, the embedding concentrates about K distinct points, representing communities.
arXiv Detail & Related papers (2021-05-03T16:36:27Z) - Hilbert-space geometry of random-matrix eigenstates [55.41644538483948]
We discuss the Hilbert-space geometry of eigenstates of parameter-dependent random-matrix ensembles.
Our results give the exact joint distribution function of the Fubini-Study metric and the Berry curvature.
We compare our results to numerical simulations of random-matrix ensembles as well as electrons in a random magnetic field.
arXiv Detail & Related papers (2020-11-06T19:00:07Z) - Statistics of the Spectral Form Factor in the Self-Dual Kicked Ising
Model [0.0]
We show that at large enough times the probability distribution agrees exactly with the prediction of Random Matrix Theory.
This behaviour is due to a recently identified additional anti-unitary symmetry of the self-dual kicked Ising model.
arXiv Detail & Related papers (2020-09-07T16:02:29Z) - Understanding Implicit Regularization in Over-Parameterized Single Index
Model [55.41685740015095]
We design regularization-free algorithms for the high-dimensional single index model.
We provide theoretical guarantees for the induced implicit regularization phenomenon.
arXiv Detail & Related papers (2020-07-16T13:27:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.