Riemannian statistics meets random matrix theory: towards learning from
high-dimensional covariance matrices
- URL: http://arxiv.org/abs/2203.00204v1
- Date: Tue, 1 Mar 2022 03:16:50 GMT
- Title: Riemannian statistics meets random matrix theory: towards learning from
high-dimensional covariance matrices
- Authors: Salem Said, Simon Heuveline, Cyrus Mostajeran
- Abstract summary: This paper shows that there seems to exist no practical method of computing the normalising factors associated with Riemannian Gaussian distributions on spaces of high-dimensional covariance matrices.
It is shown that this missing method comes from an unexpected new connection with random matrix theory.
Numerical experiments are conducted which demonstrate how this new approximation can unlock the difficulties which have impeded applications to real-world datasets.
- Score: 2.352645870795664
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Riemannian Gaussian distributions were initially introduced as basic building
blocks for learning models which aim to capture the intrinsic structure of
statistical populations of positive-definite matrices (here called covariance
matrices). While the potential applications of such models have attracted
significant attention, a major obstacle still stands in the way of these
applications: there seems to exist no practical method of computing the
normalising factors associated with Riemannian Gaussian distributions on spaces
of high-dimensional covariance matrices. The present paper shows that this
missing method comes from an unexpected new connection with random matrix
theory. Its main contribution is to prove that Riemannian Gaussian
distributions of real, complex, or quaternion covariance matrices are
equivalent to orthogonal, unitary, or symplectic log-normal matrix ensembles.
This equivalence yields a highly efficient approximation of the normalising
factors, in terms of a rather simple analytic expression. The error due to this
approximation decreases like the inverse square of dimension. Numerical
experiments are conducted which demonstrate how this new approximation can
unlock the difficulties which have impeded applications to real-world datasets
of high-dimensional covariance matrices. The paper then turns to Riemannian
Gaussian distributions of block-Toeplitz covariance matrices. These are
equivalent to yet another kind of random matrix ensembles, here called
"acosh-normal" ensembles. Orthogonal and unitary "acosh-normal" ensembles
correspond to the cases of block-Toeplitz with Toeplitz blocks, and
block-Toeplitz (with general blocks) covariance matrices, respectively.
Related papers
- Understanding Matrix Function Normalizations in Covariance Pooling through the Lens of Riemannian Geometry [63.694184882697435]
Global Covariance Pooling (GCP) has been demonstrated to improve the performance of Deep Neural Networks (DNNs) by exploiting second-order statistics of high-level representations.
arXiv Detail & Related papers (2024-07-15T07:11:44Z) - Entrywise error bounds for low-rank approximations of kernel matrices [55.524284152242096]
We derive entrywise error bounds for low-rank approximations of kernel matrices obtained using the truncated eigen-decomposition.
A key technical innovation is a delocalisation result for the eigenvectors of the kernel matrix corresponding to small eigenvalues.
We validate our theory with an empirical study of a collection of synthetic and real-world datasets.
arXiv Detail & Related papers (2024-05-23T12:26:25Z) - Learning Graphical Factor Models with Riemannian Optimization [70.13748170371889]
This paper proposes a flexible algorithmic framework for graph learning under low-rank structural constraints.
The problem is expressed as penalized maximum likelihood estimation of an elliptical distribution.
We leverage geometries of positive definite matrices and positive semi-definite matrices of fixed rank that are well suited to elliptical models.
arXiv Detail & Related papers (2022-10-21T13:19:45Z) - On confidence intervals for precision matrices and the
eigendecomposition of covariance matrices [20.20416580970697]
This paper tackles the challenge of computing confidence bounds on the individual entries of eigenvectors of a covariance matrix of fixed dimension.
We derive a method to bound the entries of the inverse covariance matrix, the so-called precision matrix.
As an application of these results, we demonstrate a new statistical test, which allows us to test for non-zero values of the precision matrix.
arXiv Detail & Related papers (2022-08-25T10:12:53Z) - Test Set Sizing Via Random Matrix Theory [91.3755431537592]
This paper uses techniques from Random Matrix Theory to find the ideal training-testing data split for a simple linear regression.
It defines "ideal" as satisfying the integrity metric, i.e. the empirical model error is the actual measurement noise.
This paper is the first to solve for the training and test size for any model in a way that is truly optimal.
arXiv Detail & Related papers (2021-12-11T13:18:33Z) - Near optimal sample complexity for matrix and tensor normal models via
geodesic convexity [5.191641077435773]
We show nonasymptotic bounds for the error achieved by the maximum likelihood estimator (MLE) in several natural metrics.
In the same regimes as our sample complexity bounds, we show that an iterative procedure to compute the MLE known as the flip-flop algorithm converges linearly with high probability.
arXiv Detail & Related papers (2021-10-14T17:47:00Z) - Riemannian Gaussian distributions, random matrix ensembles and diffusion
kernels [0.0]
We show how to compute marginals of the probability density functions on a random matrix type of symmetric spaces.
We also show how the probability density functions are a particular case of diffusion kernels of the Karlin-McGregor type, describing non-intersecting processes in the Weyl chamber of Lie groups.
arXiv Detail & Related papers (2020-11-27T11:41:29Z) - Understanding Implicit Regularization in Over-Parameterized Single Index
Model [55.41685740015095]
We design regularization-free algorithms for the high-dimensional single index model.
We provide theoretical guarantees for the induced implicit regularization phenomenon.
arXiv Detail & Related papers (2020-07-16T13:27:47Z) - Optimal Iterative Sketching with the Subsampled Randomized Hadamard
Transform [64.90148466525754]
We study the performance of iterative sketching for least-squares problems.
We show that the convergence rate for Haar and randomized Hadamard matrices are identical, andally improve upon random projections.
These techniques may be applied to other algorithms that employ randomized dimension reduction.
arXiv Detail & Related papers (2020-02-03T16:17:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.