Hyper-Laplacian Regularized Concept Factorization in Low-rank Tensor
Space for Multi-view Clustering
- URL: http://arxiv.org/abs/2304.11435v2
- Date: Tue, 1 Aug 2023 12:43:40 GMT
- Title: Hyper-Laplacian Regularized Concept Factorization in Low-rank Tensor
Space for Multi-view Clustering
- Authors: Zixiao Yu, Lele Fu, Zhiling Cai, Zhoumin Lu
- Abstract summary: We propose a hyper-Laplacian regularized concept factorization (HLRCF) in low-rank tensor space for multi-view clustering.
Specifically, we adopt the concept factorization to explore the latent cluster-wise representation of each view.
Considering that different tensor singular values associate structural information with unequal importance, we develop a self-weighted tensor Schatten p-norm.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Tensor-oriented multi-view subspace clustering has achieved significant
strides in assessing high-order correlations and improving clustering analysis
of multi-view data. Nevertheless, most of existing investigations are typically
hampered by the two flaws. First, self-representation based tensor subspace
learning usually induces high time and space complexity, and is limited in
perceiving nonlinear local structure in the embedding space. Second, the tensor
singular value decomposition (t-SVD) model redistributes each singular value
equally without considering the diverse importance among them. To well cope
with the issues, we propose a hyper-Laplacian regularized concept factorization
(HLRCF) in low-rank tensor space for multi-view clustering. Specifically, we
adopt the concept factorization to explore the latent cluster-wise
representation of each view. Further, the hypergraph Laplacian regularization
endows the model with the capability of extracting the nonlinear local
structures in the latent space. Considering that different tensor singular
values associate structural information with unequal importance, we develop a
self-weighted tensor Schatten p-norm to constrain the tensor comprised of all
cluster-wise representations. Notably, the tensor with smaller size greatly
decreases the time and space complexity in the low-rank optimization. Finally,
experimental results on eight benchmark datasets exhibit that HLRCF outperforms
other multi-view methods, showingcasing its superior performance.
Related papers
- Self-Supervised Graph Embedding Clustering [70.36328717683297]
K-means one-step dimensionality reduction clustering method has made some progress in addressing the curse of dimensionality in clustering tasks.
We propose a unified framework that integrates manifold learning with K-means, resulting in the self-supervised graph embedding framework.
arXiv Detail & Related papers (2024-09-24T08:59:51Z) - Distributional Reduction: Unifying Dimensionality Reduction and Clustering with Gromov-Wasserstein [56.62376364594194]
Unsupervised learning aims to capture the underlying structure of potentially large and high-dimensional datasets.
In this work, we revisit these approaches under the lens of optimal transport and exhibit relationships with the Gromov-Wasserstein problem.
This unveils a new general framework, called distributional reduction, that recovers DR and clustering as special cases and allows addressing them jointly within a single optimization problem.
arXiv Detail & Related papers (2024-02-03T19:00:19Z) - Multi-View Clustering via Semi-non-negative Tensor Factorization [120.87318230985653]
We develop a novel multi-view clustering based on semi-non-negative tensor factorization (Semi-NTF)
Our model directly considers the between-view relationship and exploits the between-view complementary information.
In addition, we provide an optimization algorithm for the proposed method and prove mathematically that the algorithm always converges to the stationary KKT point.
arXiv Detail & Related papers (2023-03-29T14:54:19Z) - DIFFormer: Scalable (Graph) Transformers Induced by Energy Constrained
Diffusion [66.21290235237808]
We introduce an energy constrained diffusion model which encodes a batch of instances from a dataset into evolutionary states.
We provide rigorous theory that implies closed-form optimal estimates for the pairwise diffusion strength among arbitrary instance pairs.
Experiments highlight the wide applicability of our model as a general-purpose encoder backbone with superior performance in various tasks.
arXiv Detail & Related papers (2023-01-23T15:18:54Z) - Adaptively-weighted Integral Space for Fast Multiview Clustering [54.177846260063966]
We propose an Adaptively-weighted Integral Space for Fast Multiview Clustering (AIMC) with nearly linear complexity.
Specifically, view generation models are designed to reconstruct the view observations from the latent integral space.
Experiments conducted on several realworld datasets confirm the superiority of the proposed AIMC method.
arXiv Detail & Related papers (2022-08-25T05:47:39Z) - Tensor-based Multi-view Spectral Clustering via Shared Latent Space [14.470859959783995]
Multi-view Spectral Clustering (MvSC) attracts increasing attention due to diverse data sources.
New method for MvSC is proposed via a shared latent space from the Restricted Kernel Machine framework.
arXiv Detail & Related papers (2022-07-23T17:30:54Z) - Tensor-based Intrinsic Subspace Representation Learning for Multi-view
Clustering [18.0093330816895]
We propose a novel-based Intrinsic Subspace Representation (TISRL) for multi-view clustering in this paper.
It can be seen that specific information contained in different views is fully investigated by the rank preserving decomposition.
Experimental results on nine common used real-world multi-view datasets illustrate the superiority of TISRL.
arXiv Detail & Related papers (2020-10-19T03:36:18Z) - Multi-View Spectral Clustering Tailored Tensor Low-Rank Representation [105.33409035876691]
This paper explores the problem of multi-view spectral clustering (MVSC) based on tensor low-rank modeling.
We design a novel structured tensor low-rank norm tailored to MVSC.
We show that the proposed method outperforms state-of-the-art methods to a significant extent.
arXiv Detail & Related papers (2020-04-30T11:52:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.