Learning Linear Symmetries in Data Using Moment Matching
- URL: http://arxiv.org/abs/2204.01213v1
- Date: Mon, 4 Apr 2022 02:47:37 GMT
- Title: Learning Linear Symmetries in Data Using Moment Matching
- Authors: Colin Hagemeyer
- Abstract summary: We consider the unsupervised and semi-supervised problems of learning such symmetries in a distribution directly from data.
We show that in the worst case this problem is as difficult as the graph automorphism problem.
We develop and compare theoretically and empirically the effectiveness of different methods of selecting which eigenvectors should have eigenvalue -1 in the symmetry transformation.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: It is common in machine learning and statistics to use symmetries derived
from expert knowledge to simplify problems or improve performance, using
methods like data augmentation or penalties. In this paper we consider the
unsupervised and semi-supervised problems of learning such symmetries in a
distribution directly from data in a model-free fashion. We show that in the
worst case this problem is as difficult as the graph automorphism problem.
However, if we restrict to the case where the covariance matrix has unique
eigenvalues, then the eigenvectors will also be eigenvectors of the symmetry
transformation. If we further restrict to finding orthogonal symmetries, then
the eigenvalues will be either be 1 or -1, and the problem reduces to
determining which eigenvectors are which. We develop and compare theoretically
and empirically the effectiveness of different methods of selecting which
eigenvectors should have eigenvalue -1 in the symmetry transformation, and
discuss how to extend this approach to non-orthogonal cases where we have
labels
Related papers
- Learning Infinitesimal Generators of Continuous Symmetries from Data [15.42275880523356]
We propose a novel symmetry learning algorithm based on transformations defined with one- parameter groups.
Our method is built upon minimal inductive biases, encompassing not only commonly utilized symmetries rooted in Lie groups but also extending to symmetries derived from nonlinear generators.
arXiv Detail & Related papers (2024-10-29T08:28:23Z) - Entrywise error bounds for low-rank approximations of kernel matrices [55.524284152242096]
We derive entrywise error bounds for low-rank approximations of kernel matrices obtained using the truncated eigen-decomposition.
A key technical innovation is a delocalisation result for the eigenvectors of the kernel matrix corresponding to small eigenvalues.
We validate our theory with an empirical study of a collection of synthetic and real-world datasets.
arXiv Detail & Related papers (2024-05-23T12:26:25Z) - Nonlinear SVD with Asymmetric Kernels: feature learning and asymmetric
Nystr\"om method [14.470859959783995]
Asymmetric data naturally exist in real life, such as directed graphs.
This paper tackles the asymmetric kernel-based learning problem.
Experiments show that asymmetric KSVD learns features outperforming Mercer- Kernel.
arXiv Detail & Related papers (2023-06-12T11:39:34Z) - Learning Graphical Factor Models with Riemannian Optimization [70.13748170371889]
This paper proposes a flexible algorithmic framework for graph learning under low-rank structural constraints.
The problem is expressed as penalized maximum likelihood estimation of an elliptical distribution.
We leverage geometries of positive definite matrices and positive semi-definite matrices of fixed rank that are well suited to elliptical models.
arXiv Detail & Related papers (2022-10-21T13:19:45Z) - Degeneracies and symmetry breaking in pseudo-Hermitian matrices [0.0]
We classify the eigenspace of pseudo-Hermitian matrices.
We show that symmetry breaking occurs if and only if eigenvalues of opposite kinds on the real axis of the complex eigenvalue plane.
arXiv Detail & Related papers (2022-09-14T19:18:53Z) - On confidence intervals for precision matrices and the
eigendecomposition of covariance matrices [20.20416580970697]
This paper tackles the challenge of computing confidence bounds on the individual entries of eigenvectors of a covariance matrix of fixed dimension.
We derive a method to bound the entries of the inverse covariance matrix, the so-called precision matrix.
As an application of these results, we demonstrate a new statistical test, which allows us to test for non-zero values of the precision matrix.
arXiv Detail & Related papers (2022-08-25T10:12:53Z) - When Random Tensors meet Random Matrices [50.568841545067144]
This paper studies asymmetric order-$d$ spiked tensor models with Gaussian noise.
We show that the analysis of the considered model boils down to the analysis of an equivalent spiked symmetric textitblock-wise random matrix.
arXiv Detail & Related papers (2021-12-23T04:05:01Z) - Test Set Sizing Via Random Matrix Theory [91.3755431537592]
This paper uses techniques from Random Matrix Theory to find the ideal training-testing data split for a simple linear regression.
It defines "ideal" as satisfying the integrity metric, i.e. the empirical model error is the actual measurement noise.
This paper is the first to solve for the training and test size for any model in a way that is truly optimal.
arXiv Detail & Related papers (2021-12-11T13:18:33Z) - Understanding Implicit Regularization in Over-Parameterized Single Index
Model [55.41685740015095]
We design regularization-free algorithms for the high-dimensional single index model.
We provide theoretical guarantees for the induced implicit regularization phenomenon.
arXiv Detail & Related papers (2020-07-16T13:27:47Z) - Eigendecomposition-Free Training of Deep Networks for Linear
Least-Square Problems [107.3868459697569]
We introduce an eigendecomposition-free approach to training a deep network.
We show that our approach is much more robust than explicit differentiation of the eigendecomposition.
Our method has better convergence properties and yields state-of-the-art results.
arXiv Detail & Related papers (2020-04-15T04:29:34Z) - Tackling small eigen-gaps: Fine-grained eigenvector estimation and
inference under heteroscedastic noise [28.637772416856194]
Two fundamental challenges arise in eigenvector estimation and inference for a low-rank matrix from noisy observations.
We propose estimation and uncertainty quantification procedures for an unknown eigenvector.
We establish optimal procedures to construct confidence intervals for the unknown eigenvalues.
arXiv Detail & Related papers (2020-01-14T04:26:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.