A Group Norm Regularized Factorization Model for Subspace Segmentation
- URL: http://arxiv.org/abs/2001.02568v2
- Date: Tue, 14 Jul 2020 09:13:40 GMT
- Title: A Group Norm Regularized Factorization Model for Subspace Segmentation
- Authors: Xishun Wang and Zhouwang Yang and Xingye Yue and Hui Wang
- Abstract summary: This paper proposes a group norm regularized factorization model (GNRFM) inspired by the LRR model for subspace segmentation.
Specifically, we adopt group norm regularization to make the columns of the factor matrix sparse, thereby achieving a purpose of low rank.
Compared with traditional models and algorithms, the proposed method is faster and more robust to noise, so the final clustering results are better.
- Score: 4.926716472066594
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Subspace segmentation assumes that data comes from the union of different
subspaces and the purpose of segmentation is to partition the data into the
corresponding subspace. Low-rank representation (LRR) is a classic
spectral-type method for solving subspace segmentation problems, that is, one
first obtains an affinity matrix by solving a LRR model and then performs
spectral clustering for segmentation. This paper proposes a group norm
regularized factorization model (GNRFM) inspired by the LRR model for subspace
segmentation and then designs an Accelerated Augmented Lagrangian Method (AALM)
algorithm to solve this model. Specifically, we adopt group norm regularization
to make the columns of the factor matrix sparse, thereby achieving a purpose of
low rank, which means no Singular Value Decompositions (SVD) are required and
the computational complexity of each step is greatly reduced. We obtain
affinity matrices by using different LRR models and then performing cluster
testing on different sets of synthetic noisy data and real data, respectively.
Compared with traditional models and algorithms, the proposed method is faster
and more robust to noise, so the final clustering results are better. Moreover,
the numerical results show that our algorithm converges fast and only requires
approximately ten iterations.
Related papers
- An Efficient Algorithm for Clustered Multi-Task Compressive Sensing [60.70532293880842]
Clustered multi-task compressive sensing is a hierarchical model that solves multiple compressive sensing tasks.
The existing inference algorithm for this model is computationally expensive and does not scale well in high dimensions.
We propose a new algorithm that substantially accelerates model inference by avoiding the need to explicitly compute these covariance matrices.
arXiv Detail & Related papers (2023-09-30T15:57:14Z) - Mode-wise Principal Subspace Pursuit and Matrix Spiked Covariance Model [13.082805815235975]
We introduce a novel framework called Mode-wise Principal Subspace Pursuit (MOP-UP) to extract hidden variations in both the row and column dimensions for matrix data.
The effectiveness and practical merits of the proposed framework are demonstrated through experiments on both simulated and real datasets.
arXiv Detail & Related papers (2023-07-02T13:59:47Z) - Adaptive Graph Convolutional Subspace Clustering [10.766537212211217]
Spectral-type subspace clustering algorithms have shown excellent performance in many subspace clustering applications.
In this paper, inspired by graph convolutional networks, we use the graph convolution technique to develop a feature extraction method and a coefficient matrix constraint simultaneously.
We claim that by using AGCSC, the aggregated feature representation of original data samples is suitable for subspace clustering.
arXiv Detail & Related papers (2023-05-05T10:27:23Z) - Semi-Supervised Subspace Clustering via Tensor Low-Rank Representation [64.49871502193477]
We propose a novel semi-supervised subspace clustering method, which is able to simultaneously augment the initial supervisory information and construct a discriminative affinity matrix.
Comprehensive experimental results on six commonly-used benchmark datasets demonstrate the superiority of our method over state-of-the-art methods.
arXiv Detail & Related papers (2022-05-21T01:47:17Z) - Optimal Variable Clustering for High-Dimensional Matrix Valued Data [3.1138411427556445]
We propose a new latent variable model for the features arranged in matrix form.
Under mild conditions, our algorithm attains clustering consistency in the high-dimensional setting.
We identify the optimal weight in the sense that using this weight guarantees our algorithm to be minimax rate-optimal.
arXiv Detail & Related papers (2021-12-24T02:13:04Z) - Information-Theoretic Generalization Bounds for Iterative
Semi-Supervised Learning [81.1071978288003]
In particular, we seek to understand the behaviour of the em generalization error of iterative SSL algorithms using information-theoretic principles.
Our theoretical results suggest that when the class conditional variances are not too large, the upper bound on the generalization error decreases monotonically with the number of iterations, but quickly saturates.
arXiv Detail & Related papers (2021-10-03T05:38:49Z) - Kernel Clustering with Sigmoid-based Regularization for Efficient
Segmentation of Sequential Data [3.8326963933937885]
segmentation aims at partitioning a data sequence into several non-overlapping segments that may have nonlinear and complex structures.
A popular Kernel for optimally solving this problem is dynamic programming (DP), which has quadratic computation and memory requirements.
Although many algorithms have been proposed to approximate the optimal segmentation, they have no guarantee on the quality of their solutions.
arXiv Detail & Related papers (2021-06-22T04:32:21Z) - Spatially Coherent Clustering Based on Orthogonal Nonnegative Matrix
Factorization [0.0]
We introduce in this work clustering models based on a total variation (TV) regularization procedure on the cluster membership matrix.
We provide a numerical evaluation of all proposed methods on a hyperspectral dataset obtained from a matrix-assisted laser desorption/ionisation imaging measurement.
arXiv Detail & Related papers (2021-04-25T23:40:41Z) - Multi-View Spectral Clustering with High-Order Optimal Neighborhood
Laplacian Matrix [57.11971786407279]
Multi-view spectral clustering can effectively reveal the intrinsic cluster structure among data.
This paper proposes a multi-view spectral clustering algorithm that learns a high-order optimal neighborhood Laplacian matrix.
Our proposed algorithm generates the optimal Laplacian matrix by searching the neighborhood of the linear combination of both the first-order and high-order base.
arXiv Detail & Related papers (2020-08-31T12:28:40Z) - Model Fusion with Kullback--Leibler Divergence [58.20269014662046]
We propose a method to fuse posterior distributions learned from heterogeneous datasets.
Our algorithm relies on a mean field assumption for both the fused model and the individual dataset posteriors.
arXiv Detail & Related papers (2020-07-13T03:27:45Z) - Clustering Binary Data by Application of Combinatorial Optimization
Heuristics [52.77024349608834]
We study clustering methods for binary data, first defining aggregation criteria that measure the compactness of clusters.
Five new and original methods are introduced, using neighborhoods and population behavior optimization metaheuristics.
From a set of 16 data tables generated by a quasi-Monte Carlo experiment, a comparison is performed for one of the aggregations using L1 dissimilarity, with hierarchical clustering, and a version of k-means: partitioning around medoids or PAM.
arXiv Detail & Related papers (2020-01-06T23:33:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.