Deep Embedded K-Means Clustering
- URL: http://arxiv.org/abs/2109.15149v1
- Date: Thu, 30 Sep 2021 14:12:59 GMT
- Title: Deep Embedded K-Means Clustering
- Authors: Wengang Guo, Kaiyan Lin, Wei Ye
- Abstract summary: Key idea is that representation learning and clustering can reinforce each other.
In this paper, we propose DEKM (for Deep Embedded K-Means) to answer these two questions.
Experimental results on the real-world datasets demonstrate that DEKM achieves state-of-the-art performance.
- Score: 1.5697094704362897
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recently, deep clustering methods have gained momentum because of the high
representational power of deep neural networks (DNNs) such as autoencoder. The
key idea is that representation learning and clustering can reinforce each
other: Good representations lead to good clustering while good clustering
provides good supervisory signals to representation learning. Critical
questions include: 1) How to optimize representation learning and clustering?
2) Should the reconstruction loss of autoencoder be considered always? In this
paper, we propose DEKM (for Deep Embedded K-Means) to answer these two
questions. Since the embedding space generated by autoencoder may have no
obvious cluster structures, we propose to further transform the embedding space
to a new space that reveals the cluster-structure information. This is achieved
by an orthonormal transformation matrix, which contains the eigenvectors of the
within-class scatter matrix of K-means. The eigenvalues indicate the importance
of the eigenvectors' contributions to the cluster-structure information in the
new space. Our goal is to increase the cluster-structure information. To this
end, we discard the decoder and propose a greedy method to optimize the
representation. Representation learning and clustering are alternately
optimized by DEKM. Experimental results on the real-world datasets demonstrate
that DEKM achieves state-of-the-art performance.
Related papers
- Self-Supervised Graph Embedding Clustering [70.36328717683297]
K-means one-step dimensionality reduction clustering method has made some progress in addressing the curse of dimensionality in clustering tasks.
We propose a unified framework that integrates manifold learning with K-means, resulting in the self-supervised graph embedding framework.
arXiv Detail & Related papers (2024-09-24T08:59:51Z) - MOKD: Cross-domain Finetuning for Few-shot Classification via Maximizing Optimized Kernel Dependence [97.93517982908007]
In cross-domain few-shot classification, NCC aims to learn representations to construct a metric space where few-shot classification can be performed.
In this paper, we find that there exist high similarities between NCC-learned representations of two samples from different classes.
We propose a bi-level optimization framework, emphmaximizing optimized kernel dependence (MOKD) to learn a set of class-specific representations that match the cluster structures indicated by labeled data.
arXiv Detail & Related papers (2024-05-29T05:59:52Z) - Semantic Equitable Clustering: A Simple, Fast and Effective Strategy for Vision Transformer [57.37893387775829]
We introduce a fast and balanced clustering method, named textbfSemantic textbfEquitable textbfClustering (SEC)
SEC clusters tokens based on their global semantic relevance in an efficient, straightforward manner.
We propose a versatile vision backbone, SecViT, which attains an impressive textbf84.2% image classification accuracy with only textbf27M parameters and textbf4.4G FLOPs.
arXiv Detail & Related papers (2024-05-22T04:49:00Z) - Reinforcement Graph Clustering with Unknown Cluster Number [91.4861135742095]
We propose a new deep graph clustering method termed Reinforcement Graph Clustering.
In our proposed method, cluster number determination and unsupervised representation learning are unified into a uniform framework.
In order to conduct feedback actions, the clustering-oriented reward function is proposed to enhance the cohesion of the same clusters and separate the different clusters.
arXiv Detail & Related papers (2023-08-13T18:12:28Z) - Deep Double Self-Expressive Subspace Clustering [7.875193047472789]
We propose a double self-expressive subspace clustering algorithm.
The proposed algorithm can achieve better clustering than state-of-the-art methods.
arXiv Detail & Related papers (2023-06-20T15:10:35Z) - Very Compact Clusters with Structural Regularization via Similarity and
Connectivity [3.779514860341336]
We propose an end-to-end deep clustering algorithm, i.e., Very Compact Clusters (VCC) for the general datasets.
Our proposed approach achieves better clustering performance over most of the state-of-the-art clustering methods.
arXiv Detail & Related papers (2021-06-09T23:22:03Z) - Joint Optimization of an Autoencoder for Clustering and Embedding [22.16059261437617]
We present an alternative where the autoencoder and the clustering are learned simultaneously.
That simple neural network, referred to as the clustering module, can be integrated into a deep autoencoder resulting in a deep clustering model.
arXiv Detail & Related papers (2020-12-07T14:38:10Z) - Scalable Hierarchical Agglomerative Clustering [65.66407726145619]
Existing scalable hierarchical clustering methods sacrifice quality for speed.
We present a scalable, agglomerative method for hierarchical clustering that does not sacrifice quality and scales to billions of data points.
arXiv Detail & Related papers (2020-10-22T15:58:35Z) - Structural Deep Clustering Network [45.370272344031285]
We propose a Structural Deep Clustering Network (SDCN) to integrate the structural information into deep clustering.
Specifically, we design a delivery operator to transfer the representations learned by autoencoder to the corresponding GCN layer.
In this way, the multiple structures of data, from low-order to high-order, are naturally combined with the multiple representations learned by autoencoder.
arXiv Detail & Related papers (2020-02-05T04:33:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.