AgFlow: Fast Model Selection of Penalized PCA via Implicit
Regularization Effects of Gradient Flow
- URL: http://arxiv.org/abs/2110.03273v1
- Date: Thu, 7 Oct 2021 08:57:46 GMT
- Title: AgFlow: Fast Model Selection of Penalized PCA via Implicit
Regularization Effects of Gradient Flow
- Authors: Haiyan Jiang, Haoyi Xiong, Dongrui Wu, Ji Liu, and Dejing Dou
- Abstract summary: Principal component analysis (PCA) has been widely used as an effective technique for feature extraction and dimension reduction.
In the High Dimension Low Sample Size (HDLSS) setting, one may prefer modified principal components, with penalized loadings.
We propose Approximated Gradient Flow (AgFlow) as a fast model selection method for penalized PCA.
- Score: 64.81110234990888
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Principal component analysis (PCA) has been widely used as an effective
technique for feature extraction and dimension reduction. In the High Dimension
Low Sample Size (HDLSS) setting, one may prefer modified principal components,
with penalized loadings, and automated penalty selection by implementing model
selection among these different models with varying penalties. The earlier work
[1, 2] has proposed penalized PCA, indicating the feasibility of model
selection in $L_2$- penalized PCA through the solution path of Ridge
regression, however, it is extremely time-consuming because of the intensive
calculation of matrix inverse. In this paper, we propose a fast model selection
method for penalized PCA, named Approximated Gradient Flow (AgFlow), which
lowers the computation complexity through incorporating the implicit
regularization effect introduced by (stochastic) gradient flow [3, 4] and
obtains the complete solution path of $L_2$-penalized PCA under varying
$L_2$-regularization. We perform extensive experiments on real-world datasets.
AgFlow outperforms existing methods (Oja [5], Power [6], and Shamir [7] and the
vanilla Ridge estimators) in terms of computation costs.
Related papers
- Black-Box $k$-to-$1$-PCA Reductions: Theory and Applications [19.714951004096996]
We analyze black-box deflation methods as a framework for designing $k$-PCA algorithms.
Our main contribution is significantly sharper bounds on the approximation parameter degradation of deflation methods for $k$-PCA.
We apply our framework to obtain state-of-the-art $k$-PCA algorithms robust to dataset contamination.
arXiv Detail & Related papers (2024-03-06T18:07:20Z) - Empirical Bayes Covariance Decomposition, and a solution to the Multiple
Tuning Problem in Sparse PCA [2.5382095320488673]
Sparse Principal Components Analysis (PCA) has been proposed as a way to improve both interpretability and reliability of PCA.
We present a solution to the "multiple tuning problem" using Empirical Bayes methods.
arXiv Detail & Related papers (2023-12-06T04:00:42Z) - Model-Based Reparameterization Policy Gradient Methods: Theory and
Practical Algorithms [88.74308282658133]
Reization (RP) Policy Gradient Methods (PGMs) have been widely adopted for continuous control tasks in robotics and computer graphics.
Recent studies have revealed that, when applied to long-term reinforcement learning problems, model-based RP PGMs may experience chaotic and non-smooth optimization landscapes.
We propose a spectral normalization method to mitigate the exploding variance issue caused by long model unrolls.
arXiv Detail & Related papers (2023-10-30T18:43:21Z) - Fair principal component analysis (PCA): minorization-maximization
algorithms for Fair PCA, Fair Robust PCA and Fair Sparse PCA [6.974999794070285]
We propose a new iterative algorithm to solve the fair PCA (FPCA) problem.
The proposed algorithm relies on the relaxation of a semi-orthogonality constraint which is proved to be tight at every iteration of the algorithm.
We numerically compare the performance of the proposed methods with two of the state-of-the-art approaches on synthetic data sets and a real-life data set.
arXiv Detail & Related papers (2023-05-10T08:14:32Z) - Nearly-Linear Time and Streaming Algorithms for Outlier-Robust PCA [43.106438224356175]
We develop a nearly-linear time algorithm for robust PCA with near-optimal error guarantees.
We also develop a single-pass streaming algorithm for robust PCA with memory usage nearly-linear in the dimension.
arXiv Detail & Related papers (2023-05-04T04:45:16Z) - Sparse high-dimensional linear regression with a partitioned empirical
Bayes ECM algorithm [62.997667081978825]
We propose a computationally efficient and powerful Bayesian approach for sparse high-dimensional linear regression.
Minimal prior assumptions on the parameters are used through the use of plug-in empirical Bayes estimates.
The proposed approach is implemented in the R package probe.
arXiv Detail & Related papers (2022-09-16T19:15:50Z) - Deep Equilibrium Optical Flow Estimation [80.80992684796566]
Recent state-of-the-art (SOTA) optical flow models use finite-step recurrent update operations to emulate traditional algorithms.
These RNNs impose large computation and memory overheads, and are not directly trained to model such stable estimation.
We propose deep equilibrium (DEQ) flow estimators, an approach that directly solves for the flow as the infinite-level fixed point of an implicit layer.
arXiv Detail & Related papers (2022-04-18T17:53:44Z) - Breaking the Sample Complexity Barrier to Regret-Optimal Model-Free
Reinforcement Learning [52.76230802067506]
A novel model-free algorithm is proposed to minimize regret in episodic reinforcement learning.
The proposed algorithm employs an em early-settled reference update rule, with the aid of two Q-learning sequences.
The design principle of our early-settled variance reduction method might be of independent interest to other RL settings.
arXiv Detail & Related papers (2021-10-09T21:13:48Z) - LSDAT: Low-Rank and Sparse Decomposition for Decision-based Adversarial
Attack [74.5144793386864]
LSDAT crafts perturbations in the low-dimensional subspace formed by the sparse component of the input sample and that of an adversarial sample.
LSD works directly in the image pixel domain to guarantee that non-$ell$ constraints, such as sparsity, are satisfied.
arXiv Detail & Related papers (2021-03-19T13:10:47Z) - Efficient Model-Based Collaborative Filtering with Fast Adaptive PCA [4.878057307346225]
A model-based collaborative filtering (CF) approach utilizing fast adaptive randomized singular value decomposition (SVD) is proposed.
A novel termination mechanism for the adaptive PCA is proposed to automatically determine a number of latent factors for achieving the near optimal prediction accuracy.
The proposed model-based CF approach is able to efficiently process Matlab MovieLen data with 20M ratings and exhibits more than 10X speedup over the regularized factorization based approach.
arXiv Detail & Related papers (2020-09-04T15:32:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.