Tensor Principal Component Analysis
- URL: http://arxiv.org/abs/2212.12981v2
- Date: Tue, 22 Aug 2023 15:39:36 GMT
- Title: Tensor Principal Component Analysis
- Authors: Andrii Babii, Eric Ghysels, Junsu Pan
- Abstract summary: We propose an estimation, called tensor principal component analysis (TPCA), which generalizes the traditional PCA applicable to panel data.
The TPCA algorithm and the test feature good performance in Monte Carlo experiments and are applied to sorted portfolios.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we develop new methods for analyzing high-dimensional tensor
datasets. A tensor factor model describes a high-dimensional dataset as a sum
of a low-rank component and an idiosyncratic noise, generalizing traditional
factor models for panel data. We propose an estimation algorithm, called tensor
principal component analysis (TPCA), which generalizes the traditional PCA
applicable to panel data. The algorithm involves unfolding the tensor into a
sequence of matrices along different dimensions and applying PCA to the
unfolded matrices. We provide theoretical results on the consistency and
asymptotic distribution for the TPCA estimator of loadings and factors. We also
introduce a novel test for the number of factors in a tensor factor model. The
TPCA and the test feature good performance in Monte Carlo experiments and are
applied to sorted portfolios.
Related papers
- Factor Adjusted Spectral Clustering for Mixture Models [6.806940901668607]
We propose the Factor Adjusted Spectral Clustering (FASC) algorithm, which uses an additional data denoising step via eliminating the factor component to cope with the data dependency.
Our assumption bridges many classical factor models in the literature, such as the pervasive factor model, the weak factor model, and the sparse factor model.
FASC provides significant results in many cases where traditional spectral clustering fails.
arXiv Detail & Related papers (2024-08-22T17:31:21Z) - When can weak latent factors be statistically inferred? [5.195669033269619]
This article establishes a new and comprehensive estimation and inference theory for principal component analysis (PCA)
Our theory is applicable regardless of the relative growth rate between the cross-sectional dimension $N$ and temporal dimension $TT$.
A notable technical innovation is our closed-form first-order approximation of PCA-based estimator, which paves the way for various statistical tests.
arXiv Detail & Related papers (2024-07-04T03:59:52Z) - Scaling and renormalization in high-dimensional regression [72.59731158970894]
This paper presents a succinct derivation of the training and generalization performance of a variety of high-dimensional ridge regression models.
We provide an introduction and review of recent results on these topics, aimed at readers with backgrounds in physics and deep learning.
arXiv Detail & Related papers (2024-05-01T15:59:00Z) - Tensor cumulants for statistical inference on invariant distributions [49.80012009682584]
We show that PCA becomes computationally hard at a critical value of the signal's magnitude.
We define a new set of objects, which provide an explicit, near-orthogonal basis for invariants of a given degree.
It also lets us analyze a new problem of distinguishing between different ensembles.
arXiv Detail & Related papers (2024-04-29T14:33:24Z) - Quantization of Large Language Models with an Overdetermined Basis [73.79368761182998]
We introduce an algorithm for data quantization based on the principles of Kashin representation.
Our findings demonstrate that Kashin Quantization achieves competitive or superior quality in model performance.
arXiv Detail & Related papers (2024-04-15T12:38:46Z) - Computational-Statistical Gaps in Gaussian Single-Index Models [77.1473134227844]
Single-Index Models are high-dimensional regression problems with planted structure.
We show that computationally efficient algorithms, both within the Statistical Query (SQ) and the Low-Degree Polynomial (LDP) framework, necessarily require $Omega(dkstar/2)$ samples.
arXiv Detail & Related papers (2024-03-08T18:50:19Z) - Optimal vintage factor analysis with deflation varimax [30.21039911119911]
We propose a deflation varimax procedure that solves each row of a rotation matrix sequentially.
In addition to its net computational gain and flexibility, we are able to fully guarantee for the proposed procedure in a broader context.
arXiv Detail & Related papers (2023-10-16T16:14:43Z) - Statistical Analysis of Karcher Means for Random Restricted PSD Matrices [5.867823829398135]
This paper studies an intrinsic mean model on the manifold of restricted positive semi-definite matrices and provides a non-asymptotic statistical analysis of the Karcher mean.
As an application, we show that the distributed principal component analysis algorithm, LRC-dPCA, achieves the same performance as the full sample PCA algorithm.
arXiv Detail & Related papers (2023-02-24T03:13:12Z) - Mining the Factor Zoo: Estimation of Latent Factor Models with
Sufficient Proxies [29.737081616352913]
We propose to bridge the two approaches to latent factor model estimation.
We make the latent factor model estimation robust, flexible, and statistically more accurate.
As a bonus, the number of factors is also allowed to grow.
arXiv Detail & Related papers (2022-12-25T03:10:44Z) - A Bayesian Framework on Asymmetric Mixture of Factor Analyser [0.0]
This paper introduces an MFA model with a rich and flexible class of skew normal (unrestricted) generalized hyperbolic (called SUNGH) distributions.
The SUNGH family provides considerable flexibility to model skewness in different directions as well as allowing for heavy tailed data.
Considering factor analysis models, the SUNGH family also allows for skewness and heavy tails for both the error component and factor scores.
arXiv Detail & Related papers (2022-11-01T20:19:52Z) - ER: Equivariance Regularizer for Knowledge Graph Completion [107.51609402963072]
We propose a new regularizer, namely, Equivariance Regularizer (ER)
ER can enhance the generalization ability of the model by employing the semantic equivariance between the head and tail entities.
The experimental results indicate a clear and substantial improvement over the state-of-the-art relation prediction methods.
arXiv Detail & Related papers (2022-06-24T08:18:05Z) - Generative Principal Component Analysis [47.03792476688768]
We study the problem of principal component analysis with generative modeling assumptions.
Key assumption is that the underlying signal lies near the range of an $L$-Lipschitz continuous generative model with bounded $k$-dimensional inputs.
We propose a quadratic estimator, and show that it enjoys a statistical rate of order $sqrtfracklog Lm$, where $m$ is the number of samples.
arXiv Detail & Related papers (2022-03-18T01:48:16Z) - When Random Tensors meet Random Matrices [50.568841545067144]
This paper studies asymmetric order-$d$ spiked tensor models with Gaussian noise.
We show that the analysis of the considered model boils down to the analysis of an equivalent spiked symmetric textitblock-wise random matrix.
arXiv Detail & Related papers (2021-12-23T04:05:01Z) - Equivariant vector field network for many-body system modeling [65.22203086172019]
Equivariant Vector Field Network (EVFN) is built on a novel equivariant basis and the associated scalarization and vectorization layers.
We evaluate our method on predicting trajectories of simulated Newton mechanics systems with both full and partially observed data.
arXiv Detail & Related papers (2021-10-26T14:26:25Z) - Tensor Full Feature Measure and Its Nonconvex Relaxation Applications to
Tensor Recovery [1.8899300124593645]
We propose a new tensor sparsity measure called Full Feature Measure (FFM)
It can simultaneously describe the feature dimension each dimension, and connect the Tucker rank with the tensor tube rank.
Two efficient models based on FFM are proposed, and two Alternating Multiplier Method (ADMM) algorithms are developed to solve the proposed model.
arXiv Detail & Related papers (2021-09-25T01:44:34Z) - Tensor Principal Component Analysis in High Dimensional CP Models [3.553493344868413]
We propose new algorithms for tensor CP decomposition with theoretical guarantees under mild incoherence conditions.
The composite PCA applies the principal component or singular value decompositions twice, first to a matrix unfolding of the tensor data to obtain singular vectors and then to the matrix folding of the singular vectors obtained in the first step.
We show that our implementations on synthetic data demonstrate significant practical superiority of our approach over existing methods.
arXiv Detail & Related papers (2021-08-10T03:24:32Z) - MTC: Multiresolution Tensor Completion from Partial and Coarse
Observations [49.931849672492305]
Existing completion formulation mostly relies on partial observations from a single tensor.
We propose an efficient Multi-resolution Completion model (MTC) to solve the problem.
arXiv Detail & Related papers (2021-06-14T02:20:03Z) - Stochastic Approximation for Online Tensorial Independent Component
Analysis [98.34292831923335]
Independent component analysis (ICA) has been a popular dimension reduction tool in statistical machine learning and signal processing.
In this paper, we present a by-product online tensorial algorithm that estimates for each independent component.
arXiv Detail & Related papers (2020-12-28T18:52:37Z) - Probabilistic Circuits for Variational Inference in Discrete Graphical
Models [101.28528515775842]
Inference in discrete graphical models with variational methods is difficult.
Many sampling-based methods have been proposed for estimating Evidence Lower Bound (ELBO)
We propose a new approach that leverages the tractability of probabilistic circuit models, such as Sum Product Networks (SPN)
We show that selective-SPNs are suitable as an expressive variational distribution, and prove that when the log-density of the target model is aweighted the corresponding ELBO can be computed analytically.
arXiv Detail & Related papers (2020-10-22T05:04:38Z) - Robust Tensor Principal Component Analysis: Exact Recovery via
Deterministic Model [5.414544833902815]
This paper proposes a new method to analyze Robust tensor principal component analysis (RTPCA)
It is based on the recently developed tensor-tensor product and tensor singular value decomposition (t-SVD)
arXiv Detail & Related papers (2020-08-05T16:26:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.