Supervised Linear Dimension-Reduction Methods: Review, Extensions, and
Comparisons
- URL: http://arxiv.org/abs/2109.04244v1
- Date: Thu, 9 Sep 2021 17:57:25 GMT
- Title: Supervised Linear Dimension-Reduction Methods: Review, Extensions, and
Comparisons
- Authors: Shaojie Xu, Joel Vaughan, Jie Chen, Agus Sudjianto, Vijayan Nair
- Abstract summary: Principal component analysis (PCA) is a well-known linear dimension-reduction method that has been widely used in data analysis and modeling.
This paper reviews selected techniques, extends some of them, and compares their performance through simulations.
Two of these techniques, partial least squares (PLS) and least-squares PCA (LSPCA), consistently outperform the others in this study.
- Score: 6.71092092685492
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Principal component analysis (PCA) is a well-known linear dimension-reduction
method that has been widely used in data analysis and modeling. It is an
unsupervised learning technique that identifies a suitable linear subspace for
the input variable that contains maximal variation and preserves as much
information as possible. PCA has also been used in prediction models where the
original, high-dimensional space of predictors is reduced to a smaller, more
manageable, set before conducting regression analysis. However, this approach
does not incorporate information in the response during the dimension-reduction
stage and hence can have poor predictive performance. To address this concern,
several supervised linear dimension-reduction techniques have been proposed in
the literature. This paper reviews selected techniques, extends some of them,
and compares their performance through simulations. Two of these techniques,
partial least squares (PLS) and least-squares PCA (LSPCA), consistently
outperform the others in this study.
Related papers
- Large-Scale OD Matrix Estimation with A Deep Learning Method [70.78575952309023]
The proposed method integrates deep learning and numerical optimization algorithms to infer matrix structure and guide numerical optimization.
We conducted tests to demonstrate the good generalization performance of our method on a large-scale synthetic dataset.
arXiv Detail & Related papers (2023-10-09T14:30:06Z) - Improved Privacy-Preserving PCA Using Optimized Homomorphic Matrix
Multiplication [0.0]
Principal Component Analysis (PCA) is a pivotal technique widely utilized in the realms of machine learning and data analysis.
In recent years, there have been endeavors to utilize homomorphic encryption in privacy-preserving PCA algorithms for the secure cloud computing scenario.
We propose a novel approach to privacy-preserving PCA that addresses these limitations, resulting in superior efficiency, accuracy, and scalability compared to previous approaches.
arXiv Detail & Related papers (2023-05-27T02:51:20Z) - Maximum Covariance Unfolding Regression: A Novel Covariate-based
Manifold Learning Approach for Point Cloud Data [11.34706571302446]
Point cloud data are widely used in manufacturing applications for process inspection, modeling, monitoring and optimization.
The state-of-art tensor regression techniques have effectively been used for analysis of structured point cloud data.
However, these techniques are not capable of handling unstructured point cloud data.
arXiv Detail & Related papers (2023-03-31T07:29:36Z) - An Experimental Study of Dimension Reduction Methods on Machine Learning
Algorithms with Applications to Psychometrics [77.34726150561087]
We show that dimension reduction can decrease, increase, or provide the same accuracy as no reduction of variables.
Our tentative results find that dimension reduction tends to lead to better performance when used for classification tasks.
arXiv Detail & Related papers (2022-10-19T22:07:13Z) - Sparse high-dimensional linear regression with a partitioned empirical
Bayes ECM algorithm [62.997667081978825]
We propose a computationally efficient and powerful Bayesian approach for sparse high-dimensional linear regression.
Minimal prior assumptions on the parameters are used through the use of plug-in empirical Bayes estimates.
The proposed approach is implemented in the R package probe.
arXiv Detail & Related papers (2022-09-16T19:15:50Z) - A Provably Efficient Model-Free Posterior Sampling Method for Episodic
Reinforcement Learning [50.910152564914405]
Existing posterior sampling methods for reinforcement learning are limited by being model-based or lack worst-case theoretical guarantees beyond linear MDPs.
This paper proposes a new model-free formulation of posterior sampling that applies to more general episodic reinforcement learning problems with theoretical guarantees.
arXiv Detail & Related papers (2022-08-23T12:21:01Z) - Supervised PCA: A Multiobjective Approach [70.99924195791532]
Methods for supervised principal component analysis (SPCA)
We propose a new method for SPCA that addresses both of these objectives jointly.
Our approach accommodates arbitrary supervised learning losses and, through a statistical reformulation, provides a novel low-rank extension of generalized linear models.
arXiv Detail & Related papers (2020-11-10T18:46:58Z) - Approximation Algorithms for Sparse Principal Component Analysis [57.5357874512594]
Principal component analysis (PCA) is a widely used dimension reduction technique in machine learning and statistics.
Various approaches to obtain sparse principal direction loadings have been proposed, which are termed Sparse Principal Component Analysis.
We present thresholding as a provably accurate, time, approximation algorithm for the SPCA problem.
arXiv Detail & Related papers (2020-06-23T04:25:36Z) - Physically interpretable machine learning algorithm on multidimensional
non-linear fields [0.0]
Polynomial Chaos Expansion (PCE) has long been employed as a robust representation for probabilistic input-to-output mapping.
Dimensionality Reduction (DR) techniques are increasingly used for pattern recognition and data compression.
The goal of the present paper was to combine POD and PCE for a field-measurement-based forecasting.
arXiv Detail & Related papers (2020-05-28T11:26:06Z) - Bayesian System ID: Optimal management of parameter, model, and
measurement uncertainty [0.0]
We evaluate the robustness of a probabilistic formulation of system identification (ID) to sparse, noisy, and indirect data.
We show that the log posterior has improved geometric properties compared with the objective function surfaces of traditional methods.
arXiv Detail & Related papers (2020-03-04T22:48:30Z) - A kernel Principal Component Analysis (kPCA) digest with a new backward
mapping (pre-image reconstruction) strategy [0.0]
Principal Component Analysis (PCA) is very effective if data have linear structure.
But fails in identifying a possible dimensionality reduction if data belong to a nonlinear low-dimensional manifold.
For nonlinear dimensionality reduction, kernel Principal Component Analysis (kPCA) is appreciated because of its simplicity and ease implementation.
arXiv Detail & Related papers (2020-01-07T10:30:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.