Exact and Approximation Algorithms for Sparse PCA
- URL: http://arxiv.org/abs/2008.12438v1
- Date: Fri, 28 Aug 2020 02:07:08 GMT
- Title: Exact and Approximation Algorithms for Sparse PCA
- Authors: Yongchun Li and Weijun Xie
- Abstract summary: This paper proposes two exact mixed-integer SDPs (MISDPs)
We then analyze the theoretical optimality gaps of their continuous relaxation values and prove that they are stronger than that of the state-of-art one.
Since off-the-shelf solvers, in general, have difficulty in solving MISDPs, we approximate SPCA with arbitrary accuracy by a mixed-integer linear program (MILP) of a similar size as MISDPs.
- Score: 1.7640556247739623
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Sparse PCA (SPCA) is a fundamental model in machine learning and data
analytics, which has witnessed a variety of application areas such as finance,
manufacturing, biology, healthcare. To select a prespecified-size principal
submatrix from a covariance matrix to maximize its largest eigenvalue for the
better interpretability purpose, SPCA advances the conventional PCA with both
feature selection and dimensionality reduction. This paper proposes two exact
mixed-integer SDPs (MISDPs) by exploiting the spectral decomposition of the
covariance matrix and the properties of the largest eigenvalues. We then
analyze the theoretical optimality gaps of their continuous relaxation values
and prove that they are stronger than that of the state-of-art one. We further
show that the continuous relaxations of two MISDPs can be recast as saddle
point problems without involving semi-definite cones, and thus can be
effectively solved by first-order methods such as the subgradient method. Since
off-the-shelf solvers, in general, have difficulty in solving MISDPs, we
approximate SPCA with arbitrary accuracy by a mixed-integer linear program
(MILP) of a similar size as MISDPs. To be more scalable, we also analyze greedy
and local search algorithms, prove their first-known approximation ratios, and
show that the approximation ratios are tight. Our numerical study demonstrates
that the continuous relaxation values of the proposed MISDPs are quite close to
optimality, the proposed MILP model can solve small and medium-size instances
to optimality, and the approximation algorithms work very well for all the
instances. Finally, we extend the analyses to Rank-one Sparse SVD (R1-SSVD)
with non-symmetric matrices and Sparse Fair PCA (SFPCA) when there are multiple
covariance matrices, each corresponding to a protected group.
Related papers
- Differentially Private Optimization with Sparse Gradients [60.853074897282625]
We study differentially private (DP) optimization problems under sparsity of individual gradients.
Building on this, we obtain pure- and approximate-DP algorithms with almost optimal rates for convex optimization with sparse gradients.
arXiv Detail & Related papers (2024-04-16T20:01:10Z) - Synergistic eigenanalysis of covariance and Hessian matrices for enhanced binary classification [72.77513633290056]
We present a novel approach that combines the eigenanalysis of a covariance matrix evaluated on a training set with a Hessian matrix evaluated on a deep learning model.
Our method captures intricate patterns and relationships, enhancing classification performance.
arXiv Detail & Related papers (2024-02-14T16:10:42Z) - Regularization and Variance-Weighted Regression Achieves Minimax
Optimality in Linear MDPs: Theory and Practice [79.48432795639403]
Mirror descent value iteration (MDVI) is an abstraction of Kullback-Leibler (KL) and entropy-regularized reinforcement learning (RL)
We study MDVI with linear function approximation through its sample complexity required to identify an $varepsilon$-optimal policy.
We present Variance-Weighted Least-Squares MDVI, the first theoretical algorithm that achieves nearly minimax optimal sample complexity for infinite-horizon linear MDPs.
arXiv Detail & Related papers (2023-05-22T16:13:05Z) - Support Recovery in Sparse PCA with Non-Random Missing Data [27.3669650952144]
We analyze a practical algorithm for sparse PCA on incomplete and noisy data under a general non-random sampling scheme.
We provide theoretical justification that under certain conditions, we can recover the support of the sparse leading eigenvector with high probability.
We show that our algorithm outperforms several other sparse PCA approaches especially when the observed entries have good structural properties.
arXiv Detail & Related papers (2023-02-03T04:20:25Z) - Asymptotically Unbiased Instance-wise Regularized Partial AUC
Optimization: Theory and Algorithm [101.44676036551537]
One-way Partial AUC (OPAUC) and Two-way Partial AUC (TPAUC) measures the average performance of a binary classifier.
Most of the existing methods could only optimize PAUC approximately, leading to inevitable biases that are not controllable.
We present a simpler reformulation of the PAUC problem via distributional robust optimization AUC.
arXiv Detail & Related papers (2022-10-08T08:26:22Z) - Jointly Modeling and Clustering Tensors in High Dimensions [6.072664839782975]
We consider the problem of jointly benchmarking and clustering of tensors.
We propose an efficient high-maximization algorithm that converges geometrically to a neighborhood that is within statistical precision.
arXiv Detail & Related papers (2021-04-15T21:06:16Z) - Simple and Near-Optimal MAP Inference for Nonsymmetric DPPs [3.3504365823045044]
We study the problem of maximum a posteriori (MAP) inference for determinantal point processes defined by a nonsymmetric kernel matrix.
We obtain the first multiplicative approximation guarantee for this problem using local search.
Our approximation factor of $kO(k)$ is nearly tight, and we show theoretically and experimentally that it compares favorably to the state-of-the-art methods for this problem.
arXiv Detail & Related papers (2021-02-10T09:34:44Z) - Approximation Algorithms for Sparse Principal Component Analysis [57.5357874512594]
Principal component analysis (PCA) is a widely used dimension reduction technique in machine learning and statistics.
Various approaches to obtain sparse principal direction loadings have been proposed, which are termed Sparse Principal Component Analysis.
We present thresholding as a provably accurate, time, approximation algorithm for the SPCA problem.
arXiv Detail & Related papers (2020-06-23T04:25:36Z) - A Dynamical Systems Approach for Convergence of the Bayesian EM
Algorithm [59.99439951055238]
We show how (discrete-time) Lyapunov stability theory can serve as a powerful tool to aid, or even lead, in the analysis (and potential design) of optimization algorithms that are not necessarily gradient-based.
The particular ML problem that this paper focuses on is that of parameter estimation in an incomplete-data Bayesian framework via the popular optimization algorithm known as maximum a posteriori expectation-maximization (MAP-EM)
We show that fast convergence (linear or quadratic) is achieved, which could have been difficult to unveil without our adopted S&C approach.
arXiv Detail & Related papers (2020-06-23T01:34:18Z) - Best Principal Submatrix Selection for the Maximum Entropy Sampling
Problem: Scalable Algorithms and Performance Guarantees [1.7640556247739623]
MESP has been widely applied to many areas, including healthcare, power system, manufacturing and data science.
We derive a novel convex integer program for MESP and show that its continuous relaxation yields a near-optimal solution.
Our numerical experiments demonstrate that these approximation algorithms can efficiently solve medium-sized and large-scale instances to near-optimality.
arXiv Detail & Related papers (2020-01-23T14:14:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.