Generalized Clustering and Multi-Manifold Learning with Geometric
Structure Preservation
- URL: http://arxiv.org/abs/2009.09590v4
- Date: Sat, 9 Oct 2021 02:53:38 GMT
- Title: Generalized Clustering and Multi-Manifold Learning with Geometric
Structure Preservation
- Authors: Lirong Wu, Zicheng Liu, Zelin Zang, Jun Xia, Siyuan Li, Stan. Z Li
- Abstract summary: We propose a novel Generalized Clustering and Multi-manifold Learning (GCML) framework with geometric structure preservation for generalized data.
In the proposed framework, manifold clustering is done in the latent space guided by a clustering loss.
To overcome the problem that the clustering-oriented loss may deteriorate the geometric structure of the latent space, an isometric loss is proposed for preserving intra-manifold structure locally and a ranking loss for inter-manifold structure globally.
- Score: 47.65743823937763
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Though manifold-based clustering has become a popular research topic, we
observe that one important factor has been omitted by these works, namely that
the defined clustering loss may corrupt the local and global structure of the
latent space. In this paper, we propose a novel Generalized Clustering and
Multi-manifold Learning (GCML) framework with geometric structure preservation
for generalized data, i.e., not limited to 2-D image data and has a wide range
of applications in speech, text, and biology domains. In the proposed
framework, manifold clustering is done in the latent space guided by a
clustering loss. To overcome the problem that the clustering-oriented loss may
deteriorate the geometric structure of the latent space, an isometric loss is
proposed for preserving intra-manifold structure locally and a ranking loss for
inter-manifold structure globally. Extensive experimental results have shown
that GCML exhibits superior performance to counterparts in terms of qualitative
visualizations and quantitative metrics, which demonstrates the effectiveness
of preserving geometric structure.
Related papers
- ClusterGraph: a new tool for visualization and compression of multidimensional data [0.0]
This paper provides an additional layer on the output of any clustering algorithm.
It provides information about the global layout of clusters, obtained from the considered clustering algorithm.
arXiv Detail & Related papers (2024-11-08T09:40:54Z) - Distributional Reduction: Unifying Dimensionality Reduction and Clustering with Gromov-Wasserstein [56.62376364594194]
Unsupervised learning aims to capture the underlying structure of potentially large and high-dimensional datasets.
In this work, we revisit these approaches under the lens of optimal transport and exhibit relationships with the Gromov-Wasserstein problem.
This unveils a new general framework, called distributional reduction, that recovers DR and clustering as special cases and allows addressing them jointly within a single optimization problem.
arXiv Detail & Related papers (2024-02-03T19:00:19Z) - Scalable manifold learning by uniform landmark sampling and constrained
locally linear embedding [0.6144680854063939]
We propose a scalable manifold learning (scML) method that can manipulate large-scale and high-dimensional data in an efficient manner.
We empirically validated the effectiveness of scML on synthetic datasets and real-world benchmarks of different types.
scML scales well with increasing data sizes and embedding dimensions, and exhibits promising performance in preserving the global structure.
arXiv Detail & Related papers (2024-01-02T08:43:06Z) - Efficient Multi-View Graph Clustering with Local and Global Structure
Preservation [59.49018175496533]
We propose a novel anchor-based multi-view graph clustering framework termed Efficient Multi-View Graph Clustering with Local and Global Structure Preservation (EMVGC-LG)
Specifically, EMVGC-LG jointly optimize anchor construction and graph learning to enhance the clustering quality.
In addition, EMVGC-LG inherits the linear complexity of existing AMVGC methods respecting the sample number.
arXiv Detail & Related papers (2023-08-31T12:12:30Z) - Local and Global Structure Preservation Based Spectral Clustering [1.8047694351309205]
This paper extends Spectral Clustering to Local and Global Structure Preservation Based Spectral Clustering (LGPSC)
LGPSC incorporates both global structure and local neighborhood structure simultaneously.
arXiv Detail & Related papers (2022-10-23T16:36:05Z) - Enhancing cluster analysis via topological manifold learning [0.3823356975862006]
We show that inferring the topological structure of a dataset before clustering can considerably enhance cluster detection.
We combine manifold learning method UMAP for inferring the topological structure with density-based clustering method DBSCAN.
arXiv Detail & Related papers (2022-07-01T15:53:39Z) - Structure-Aware Feature Generation for Zero-Shot Learning [108.76968151682621]
We introduce a novel structure-aware feature generation scheme, termed as SA-GAN, to account for the topological structure in learning both the latent space and the generative networks.
Our method significantly enhances the generalization capability on unseen-classes and consequently improve the classification performance.
arXiv Detail & Related papers (2021-08-16T11:52:08Z) - Unveiling the Potential of Structure-Preserving for Weakly Supervised
Object Localization [71.79436685992128]
We propose a two-stage approach, termed structure-preserving activation (SPA), towards fully leveraging the structure information incorporated in convolutional features for WSOL.
In the first stage, a restricted activation module (RAM) is designed to alleviate the structure-missing issue caused by the classification network.
In the second stage, we propose a post-process approach, termed self-correlation map generating (SCG) module to obtain structure-preserving localization maps.
arXiv Detail & Related papers (2021-03-08T03:04:14Z) - Scalable Hierarchical Agglomerative Clustering [65.66407726145619]
Existing scalable hierarchical clustering methods sacrifice quality for speed.
We present a scalable, agglomerative method for hierarchical clustering that does not sacrifice quality and scales to billions of data points.
arXiv Detail & Related papers (2020-10-22T15:58:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.