Optimal vintage factor analysis with deflation varimax
- URL: http://arxiv.org/abs/2310.10545v1
- Date: Mon, 16 Oct 2023 16:14:43 GMT
- Title: Optimal vintage factor analysis with deflation varimax
- Authors: Xin Bing, Dian Jin and Yuqian Zhang
- Abstract summary: Vintage factor analysis aims to first find a low-dimensional representation of the original data, and then to seek a such that the rotated low-dimensional representation is scientifically meaningful.
Perhaps most widely used vintage factor analysis is Principal Component Analysis (PCA) followed by varimax representation.
In this paper, we propose a deflation-to-optimization procedure that solves each row matrix sequentially.
- Score: 18.50195604586597
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Vintage factor analysis is one important type of factor analysis that aims to
first find a low-dimensional representation of the original data, and then to
seek a rotation such that the rotated low-dimensional representation is
scientifically meaningful. Perhaps the most widely used vintage factor analysis
is the Principal Component Analysis (PCA) followed by the varimax rotation.
Despite its popularity, little theoretical guarantee can be provided mainly
because varimax rotation requires to solve a non-convex optimization over the
set of orthogonal matrices.
In this paper, we propose a deflation varimax procedure that solves each row
of an orthogonal matrix sequentially. In addition to its net computational gain
and flexibility, we are able to fully establish theoretical guarantees for the
proposed procedure in a broad context.
Adopting this new varimax approach as the second step after PCA, we further
analyze this two step procedure under a general class of factor models. Our
results show that it estimates the factor loading matrix in the optimal rate
when the signal-to-noise-ratio (SNR) is moderate or large. In the low SNR
regime, we offer possible improvement over using PCA and the deflation
procedure when the additive noise under the factor model is structured. The
modified procedure is shown to be optimal in all SNR regimes. Our theory is
valid for finite sample and allows the number of the latent factors to grow
with the sample size as well as the ambient dimension to grow with, or even
exceed, the sample size.
Extensive simulation and real data analysis further corroborate our
theoretical findings.
Related papers
- Revisit CP Tensor Decomposition: Statistical Optimality and Fast Convergence [6.724750970258851]
We revisit Canonical Polyadic (CP) tensor decomposition from a statistical perspective.<n>We provide a comprehensive theoretical analysis of Alternating Least Squares (ALS) under a signal-plus-noise model.
arXiv Detail & Related papers (2025-05-29T03:42:03Z) - T-Rex: Fitting a Robust Factor Model via Expectation-Maximization [0.0]
We propose a novel expectation-maximization (EM) algorithm for robustly fitting statistical factor models.<n>Our approach is based on Tyler's M-estimator of the scatter matrix for an elliptical distribution.<n>We present numerical experiments on both synthetic and real examples.
arXiv Detail & Related papers (2025-05-17T18:53:06Z) - On the Asymptotic Mean Square Error Optimality of Diffusion Models [10.72484143420088]
Diffusion models (DMs) as generative priors have recently shown great potential for denoising tasks.
This paper proposes a novel denoising strategy inspired by the structure of the MSE-optimal conditional mean (CME)
The resulting DM-based denoiser can be conveniently employed using a pre-trained DM, being particularly fast by truncating reverse diffusion steps.
arXiv Detail & Related papers (2024-03-05T13:25:44Z) - Probabilistic Unrolling: Scalable, Inverse-Free Maximum Likelihood
Estimation for Latent Gaussian Models [69.22568644711113]
We introduce probabilistic unrolling, a method that combines Monte Carlo sampling with iterative linear solvers to circumvent matrix inversions.
Our theoretical analyses reveal that unrolling and backpropagation through the iterations of the solver can accelerate gradient estimation for maximum likelihood estimation.
In experiments on simulated and real data, we demonstrate that probabilistic unrolling learns latent Gaussian models up to an order of magnitude faster than gradient EM, with minimal losses in model performance.
arXiv Detail & Related papers (2023-06-05T21:08:34Z) - Regularization and Variance-Weighted Regression Achieves Minimax
Optimality in Linear MDPs: Theory and Practice [79.48432795639403]
Mirror descent value iteration (MDVI) is an abstraction of Kullback-Leibler (KL) and entropy-regularized reinforcement learning (RL)
We study MDVI with linear function approximation through its sample complexity required to identify an $varepsilon$-optimal policy.
We present Variance-Weighted Least-Squares MDVI, the first theoretical algorithm that achieves nearly minimax optimal sample complexity for infinite-horizon linear MDPs.
arXiv Detail & Related papers (2023-05-22T16:13:05Z) - On High dimensional Poisson models with measurement error: hypothesis
testing for nonlinear nonconvex optimization [13.369004892264146]
We estimation and testing regression model with high dimensionals, which has wide applications in analyzing data.
We propose to estimate regression parameter through minimizing penalized consistency.
The proposed method is applied to the Alzheimer's Disease Initiative.
arXiv Detail & Related papers (2022-12-31T06:58:42Z) - Structured Optimal Variational Inference for Dynamic Latent Space Models [16.531262817315696]
We consider a latent space model for dynamic networks, where our objective is to estimate the pairwise inner products plus the intercept of the latent positions.
To balance posterior inference and computational scalability, we consider a structured mean-field variational inference framework.
arXiv Detail & Related papers (2022-09-29T22:10:42Z) - Sparse high-dimensional linear regression with a partitioned empirical
Bayes ECM algorithm [62.997667081978825]
We propose a computationally efficient and powerful Bayesian approach for sparse high-dimensional linear regression.
Minimal prior assumptions on the parameters are used through the use of plug-in empirical Bayes estimates.
The proposed approach is implemented in the R package probe.
arXiv Detail & Related papers (2022-09-16T19:15:50Z) - Inverting brain grey matter models with likelihood-free inference: a
tool for trustable cytoarchitecture measurements [62.997667081978825]
characterisation of the brain grey matter cytoarchitecture with quantitative sensitivity to soma density and volume remains an unsolved challenge in dMRI.
We propose a new forward model, specifically a new system of equations, requiring a few relatively sparse b-shells.
We then apply modern tools from Bayesian analysis known as likelihood-free inference (LFI) to invert our proposed model.
arXiv Detail & Related papers (2021-11-15T09:08:27Z) - Generalized Matrix Factorization: efficient algorithms for fitting
generalized linear latent variable models to large data arrays [62.997667081978825]
Generalized Linear Latent Variable models (GLLVMs) generalize such factor models to non-Gaussian responses.
Current algorithms for estimating model parameters in GLLVMs require intensive computation and do not scale to large datasets.
We propose a new approach for fitting GLLVMs to high-dimensional datasets, based on approximating the model using penalized quasi-likelihood.
arXiv Detail & Related papers (2020-10-06T04:28:19Z) - On the minmax regret for statistical manifolds: the role of curvature [68.8204255655161]
Two-part codes and the minimum description length have been successful in delivering procedures to single out the best models.
We derive a sharper expression than the standard one given by the complexity, where the scalar curvature of the Fisher information metric plays a dominant role.
arXiv Detail & Related papers (2020-07-06T17:28:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.