Angular Embedding: A New Angular Robust Principal Component Analysis
- URL: http://arxiv.org/abs/2011.11013v1
- Date: Sun, 22 Nov 2020 13:36:56 GMT
- Title: Angular Embedding: A New Angular Robust Principal Component Analysis
- Authors: Shenglan Liu, Yang Yu
- Abstract summary: It is a serious problem that principal component analysis is sensitive to outliers.
The existing state-of-the-art RPCA approaches cannot easily remove or tolerate outliers by a non-iterative manner.
This paper proposes Angular Embedding (AE) to formulate a straightforward RPCA approach based on angular density.
Furthermore, a trimmed AE (TAE) is introduced to deal with data with large scale outliers.
- Score: 10.120548476934186
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: As a widely used method in machine learning, principal component analysis
(PCA) shows excellent properties for dimensionality reduction. It is a serious
problem that PCA is sensitive to outliers, which has been improved by numerous
Robust PCA (RPCA) versions. However, the existing state-of-the-art RPCA
approaches cannot easily remove or tolerate outliers by a non-iterative manner.
To tackle this issue, this paper proposes Angular Embedding (AE) to formulate a
straightforward RPCA approach based on angular density, which is improved for
large scale or high-dimensional data. Furthermore, a trimmed AE (TAE) is
introduced to deal with data with large scale outliers. Extensive experiments
on both synthetic and real-world datasets with vector-level or pixel-level
outliers demonstrate that the proposed AE/TAE outperforms the state-of-the-art
RPCA based methods.
Related papers
- Small Object Detection via Coarse-to-fine Proposal Generation and
Imitation Learning [52.06176253457522]
We propose a two-stage framework tailored for small object detection based on the Coarse-to-fine pipeline and Feature Imitation learning.
CFINet achieves state-of-the-art performance on the large-scale small object detection benchmarks, SODA-D and SODA-A.
arXiv Detail & Related papers (2023-08-18T13:13:09Z) - PCA-Boosted Autoencoders for Nonlinear Dimensionality Reduction in Low
Data Regimes [0.2925461470287228]
We propose a technique that harnesses the best of both worlds: an autoencoder that leverages PCA to perform well on scarce nonlinear data.
A synthetic example is presented first to study the effects of data nonlinearity and size on the performance of the proposed method.
arXiv Detail & Related papers (2022-05-23T23:46:52Z) - Learned Robust PCA: A Scalable Deep Unfolding Approach for
High-Dimensional Outlier Detection [23.687598836093333]
Robust principal component analysis is a critical tool in machine learning, which detects outliers in the task of low-rank reconstruction.
In this paper, we propose a scalable and learnable approach for high-dimensional RPCA problems which we call LRPCA.
We show that LRPCA outperforms the state-of-the-art RPCA algorithms, such as neurald AltProj, on both datasets real-world applications.
arXiv Detail & Related papers (2021-10-11T23:37:55Z) - AgFlow: Fast Model Selection of Penalized PCA via Implicit
Regularization Effects of Gradient Flow [64.81110234990888]
Principal component analysis (PCA) has been widely used as an effective technique for feature extraction and dimension reduction.
In the High Dimension Low Sample Size (HDLSS) setting, one may prefer modified principal components, with penalized loadings.
We propose Approximated Gradient Flow (AgFlow) as a fast model selection method for penalized PCA.
arXiv Detail & Related papers (2021-10-07T08:57:46Z) - FAST-PCA: A Fast and Exact Algorithm for Distributed Principal Component
Analysis [12.91948651812873]
Principal Component Analysis (PCA) is a fundamental data preprocessing tool in the world of machine learning.
This paper proposes a distributed PCA algorithm called FAST-PCA (Fast and exAct diSTributed PCA)
arXiv Detail & Related papers (2021-08-27T16:10:59Z) - Enhanced Principal Component Analysis under A Collaborative-Robust
Framework [89.28334359066258]
We introduce a general collaborative-robust weight learning framework that combines weight learning and robust loss in a non-trivial way.
Under the proposed framework, only a part of well-fitting samples are activated which indicates more importance during training, and others, whose errors are large, will not be ignored.
In particular, the negative effects of inactivated samples are alleviated by the robust loss function.
arXiv Detail & Related papers (2021-03-22T15:17:37Z) - Supervised PCA: A Multiobjective Approach [70.99924195791532]
Methods for supervised principal component analysis (SPCA)
We propose a new method for SPCA that addresses both of these objectives jointly.
Our approach accommodates arbitrary supervised learning losses and, through a statistical reformulation, provides a novel low-rank extension of generalized linear models.
arXiv Detail & Related papers (2020-11-10T18:46:58Z) - Improved Dimensionality Reduction of various Datasets using Novel
Multiplicative Factoring Principal Component Analysis (MPCA) [0.0]
We present an improvement to the traditional PCA approach called Multiplicative factoring Principal Component Analysis.
The advantage of MPCA over the traditional PCA is that a penalty is imposed on the occurrence space through a multiplier to make negligible the effect of outliers in seeking out projections.
arXiv Detail & Related papers (2020-09-25T12:30:15Z) - Principal Ellipsoid Analysis (PEA): Efficient non-linear dimension
reduction & clustering [9.042239247913642]
This article focuses on improving upon PCA and k-means, by allowing nonlinear relations in the data and more flexible cluster shapes.
The key contribution is a new framework for Principal Analysis (PEA), defining a simple and computationally efficient alternative to PCA.
In a rich variety of real data clustering applications, PEA is shown to do as well as k-means for simple datasets, while dramatically improving performance in more complex settings.
arXiv Detail & Related papers (2020-08-17T06:25:50Z) - Approximation Algorithms for Sparse Principal Component Analysis [57.5357874512594]
Principal component analysis (PCA) is a widely used dimension reduction technique in machine learning and statistics.
Various approaches to obtain sparse principal direction loadings have been proposed, which are termed Sparse Principal Component Analysis.
We present thresholding as a provably accurate, time, approximation algorithm for the SPCA problem.
arXiv Detail & Related papers (2020-06-23T04:25:36Z) - Repulsive Mixture Models of Exponential Family PCA for Clustering [127.90219303669006]
The mixture extension of exponential family principal component analysis ( EPCA) was designed to encode much more structural information about data distribution than the traditional EPCA.
The traditional mixture of local EPCAs has the problem of model redundancy, i.e., overlaps among mixing components, which may cause ambiguity for data clustering.
In this paper, a repulsiveness-encouraging prior is introduced among mixing components and a diversified EPCA mixture (DEPCAM) model is developed in the Bayesian framework.
arXiv Detail & Related papers (2020-04-07T04:07:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.