Kernel Two-Dimensional Ridge Regression for Subspace Clustering
- URL: http://arxiv.org/abs/2011.01477v1
- Date: Tue, 3 Nov 2020 04:52:46 GMT
- Title: Kernel Two-Dimensional Ridge Regression for Subspace Clustering
- Authors: Chong Peng, Qian Zhang, Zhao Kang, Chenglizhao Chen, and Qiang Cheng
- Abstract summary: We propose a novel subspace clustering method for 2D data.
It directly uses 2D data as inputs such that the learning of representations benefits from inherent structures and relationships of the data.
- Score: 45.651770340521786
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Subspace clustering methods have been widely studied recently. When the
inputs are 2-dimensional (2D) data, existing subspace clustering methods
usually convert them into vectors, which severely damages inherent structures
and relationships from original data. In this paper, we propose a novel
subspace clustering method for 2D data. It directly uses 2D data as inputs such
that the learning of representations benefits from inherent structures and
relationships of the data. It simultaneously seeks image projection and
representation coefficients such that they mutually enhance each other and lead
to powerful data representations. An efficient algorithm is developed to solve
the proposed objective function with provable decreasing and convergence
property. Extensive experimental results verify the effectiveness of the new
method.
Related papers
- Distributional Reduction: Unifying Dimensionality Reduction and Clustering with Gromov-Wasserstein [56.62376364594194]
Unsupervised learning aims to capture the underlying structure of potentially large and high-dimensional datasets.
In this work, we revisit these approaches under the lens of optimal transport and exhibit relationships with the Gromov-Wasserstein problem.
This unveils a new general framework, called distributional reduction, that recovers DR and clustering as special cases and allows addressing them jointly within a single optimization problem.
arXiv Detail & Related papers (2024-02-03T19:00:19Z) - Learning A Disentangling Representation For PU Learning [18.94726971543125]
We propose to learn a neural network-based data representation using a loss function that can be used to project the unlabeled data into two clusters.
We conduct experiments on simulated PU data that demonstrate the improved performance of our proposed method compared to the current state-of-the-art approaches.
arXiv Detail & Related papers (2023-10-05T18:33:32Z) - Improved Distribution Matching for Dataset Condensation [91.55972945798531]
We propose a novel dataset condensation method based on distribution matching.
Our simple yet effective method outperforms most previous optimization-oriented methods with much fewer computational resources.
arXiv Detail & Related papers (2023-07-19T04:07:33Z) - Index $t$-SNE: Tracking Dynamics of High-Dimensional Datasets with
Coherent Embeddings [1.7188280334580195]
This paper presents a methodology to reuse an embedding to create a new one, where cluster positions are preserved.
The proposed algorithm has the same complexity as the original $t$-SNE to embed new items, and a lower one when considering the embedding of a dataset sliced into sub-pieces.
arXiv Detail & Related papers (2021-09-22T06:45:37Z) - Graph Constrained Data Representation Learning for Human Motion
Segmentation [14.611777974037194]
We propose a novel unsupervised model that learns a representation of the data and digs clustering information from the data itself.
Experimental results on four benchmark datasets for HMS demonstrate that our approach achieves significantly better clustering performance then state-of-the-art methods.
arXiv Detail & Related papers (2021-07-28T13:49:16Z) - A Local Similarity-Preserving Framework for Nonlinear Dimensionality
Reduction with Neural Networks [56.068488417457935]
We propose a novel local nonlinear approach named Vec2vec for general purpose dimensionality reduction.
To train the neural network, we build the neighborhood similarity graph of a matrix and define the context of data points.
Experiments of data classification and clustering on eight real datasets show that Vec2vec is better than several classical dimensionality reduction methods in the statistical hypothesis test.
arXiv Detail & Related papers (2021-03-10T23:10:47Z) - Manifold Partition Discriminant Analysis [42.11470531267327]
We propose a novel algorithm for supervised dimensionality reduction named Manifold Partition Discriminant Analysis (MPDA)
It aims to find a linear embedding space where the within-class similarity is achieved along the direction that is consistent with the local variation of the data manifold.
MPDA explicitly parameterizes the connections of tangent spaces and represents the data manifold in a piecewise manner.
arXiv Detail & Related papers (2020-11-23T16:33:23Z) - Two-Dimensional Semi-Nonnegative Matrix Factorization for Clustering [50.43424130281065]
We propose a new Semi-Nonnegative Matrix Factorization method for 2-dimensional (2D) data, named TS-NMF.
It overcomes the drawback of existing methods that seriously damage the spatial information of the data by converting 2D data to vectors in a preprocessing step.
arXiv Detail & Related papers (2020-05-19T05:54:14Z) - Ellipsoidal Subspace Support Vector Data Description [98.67884574313292]
We propose a novel method for transforming data into a low-dimensional space optimized for one-class classification.
We provide both linear and non-linear formulations for the proposed method.
The proposed method is noticed to converge much faster than recently proposed Subspace Support Vector Data Description.
arXiv Detail & Related papers (2020-03-20T21:31:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.