Spectral Learning on Matrices and Tensors
- URL: http://arxiv.org/abs/2004.07984v1
- Date: Thu, 16 Apr 2020 22:53:00 GMT
- Title: Spectral Learning on Matrices and Tensors
- Authors: Majid Janzamin, Rong Ge, Jean Kossaifi and Anima Anandkumar
- Abstract summary: We show that tensor decomposition can pick up latent effects that are missed by matrix methods.
We also outline computational techniques to design efficient tensor decomposition methods.
- Score: 74.88243719463053
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Spectral methods have been the mainstay in several domains such as machine
learning and scientific computing. They involve finding a certain kind of
spectral decomposition to obtain basis functions that can capture important
structures for the problem at hand. The most common spectral method is the
principal component analysis (PCA). It utilizes the top eigenvectors of the
data covariance matrix, e.g. to carry out dimensionality reduction. This data
pre-processing step is often effective in separating signal from noise. PCA and
other spectral techniques applied to matrices have several limitations. By
limiting to only pairwise moments, they are effectively making a Gaussian
approximation on the underlying data and fail on data with hidden variables
which lead to non-Gaussianity. However, in most data sets, there are latent
effects that cannot be directly observed, e.g., topics in a document corpus, or
underlying causes of a disease. By extending the spectral decomposition methods
to higher order moments, we demonstrate the ability to learn a wide range of
latent variable models efficiently. Higher-order moments can be represented by
tensors, and intuitively, they can encode more information than just pairwise
moment matrices. More crucially, tensor decomposition can pick up latent
effects that are missed by matrix methods, e.g. uniquely identify
non-orthogonal components. Exploiting these aspects turns out to be fruitful
for provable unsupervised learning of a wide range of latent variable models.
We also outline the computational techniques to design efficient tensor
decomposition methods. We introduce Tensorly, which has a simple python
interface for expressing tensor operations. It has a flexible back-end system
supporting NumPy, PyTorch, TensorFlow and MXNet amongst others, allowing
multi-GPU and CPU operations and seamless integration with deep-learning
functionalities.
Related papers
- Tensor cumulants for statistical inference on invariant distributions [49.80012009682584]
We show that PCA becomes computationally hard at a critical value of the signal's magnitude.
We define a new set of objects, which provide an explicit, near-orthogonal basis for invariants of a given degree.
It also lets us analyze a new problem of distinguishing between different ensembles.
arXiv Detail & Related papers (2024-04-29T14:33:24Z) - Laplacian Canonization: A Minimalist Approach to Sign and Basis
Invariant Spectral Embedding [36.61907023057978]
Spectral embedding is a powerful graph computation technique that has received a lot of attention recently due to its effectiveness on Graph Transformers.
Previous methods developed costly approaches to learn new invariants and suffer from high complexity.
In this work, we explore a minimal approach that resolves the ambiguity issues by directly finding canonical directions for the eigenvectors.
arXiv Detail & Related papers (2023-10-28T14:35:10Z) - Stochastic Parallelizable Eigengap Dilation for Large Graph Clustering [12.544602297450533]
We aim to identify clusters of nodes where most edges fall within clusters and only few edges fall between clusters.
A core step of spectral clustering is performing an eigendecomposition of the corresponding graph Laplacian matrix.
This paper introduces a parallelizable approach to dilating the spectrum in order to accelerate SVD solvers and in turn, spectral clustering.
arXiv Detail & Related papers (2022-07-29T10:13:07Z) - Softmax-free Linear Transformers [90.83157268265654]
Vision transformers (ViTs) have pushed the state-of-the-art for visual perception tasks.
Existing methods are either theoretically flawed or empirically ineffective for visual recognition.
We propose a family of Softmax-Free Transformers (SOFT)
arXiv Detail & Related papers (2022-07-05T03:08:27Z) - Probabilistic Simplex Component Analysis [66.30587591100566]
PRISM is a probabilistic simplex component analysis approach to identifying the vertices of a data-circumscribing simplex from data.
The problem has a rich variety of applications, the most notable being hyperspectral unmixing in remote sensing and non-negative matrix factorization in machine learning.
arXiv Detail & Related papers (2021-03-18T05:39:00Z) - Spectral Methods for Data Science: A Statistical Perspective [37.2486912080998]
Spectral methods have emerged as a simple yet surprisingly effective approach for extracting information from massive, noisy and incomplete data.
This book aims to present a systematic, comprehensive, yet accessible introduction to spectral methods from a modern statistical perspective.
arXiv Detail & Related papers (2020-12-15T18:40:56Z) - Learnable Graph-regularization for Matrix Decomposition [5.9394103049943485]
We propose a learnable graph-regularization model for matrix decomposition.
It builds a bridge between graph-regularized methods and probabilistic matrix decomposition models.
It learns two graphical structures in real-time in an iterative manner via sparse precision matrix estimation.
arXiv Detail & Related papers (2020-10-16T17:12:39Z) - Linear-Sample Learning of Low-Rank Distributions [56.59844655107251]
We show that learning $ktimes k$, rank-$r$, matrices to normalized $L_1$ distance requires $Omega(frackrepsilon2)$ samples.
We propose an algorithm that uses $cal O(frackrepsilon2log2fracepsilon)$ samples, a number linear in the high dimension, and nearly linear in the matrices, typically low, rank proofs.
arXiv Detail & Related papers (2020-09-30T19:10:32Z) - A Solution for Large Scale Nonlinear Regression with High Rank and
Degree at Constant Memory Complexity via Latent Tensor Reconstruction [0.0]
This paper proposes a novel method for learning highly nonlinear, multivariate functions from examples.
Our method takes advantage of the property that continuous functions can be approximated by bys, which in turn are representable by tensors.
For learning the models, we present an efficient-based algorithm that can be implemented in linear time.
arXiv Detail & Related papers (2020-05-04T14:49:14Z) - Improved guarantees and a multiple-descent curve for Column Subset
Selection and the Nystr\"om method [76.73096213472897]
We develop techniques which exploit spectral properties of the data matrix to obtain improved approximation guarantees.
Our approach leads to significantly better bounds for datasets with known rates of singular value decay.
We show that both our improved bounds and the multiple-descent curve can be observed on real datasets simply by varying the RBF parameter.
arXiv Detail & Related papers (2020-02-21T00:43:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.