Learning the Precise Feature for Cluster Assignment
- URL: http://arxiv.org/abs/2106.06159v1
- Date: Fri, 11 Jun 2021 04:08:54 GMT
- Title: Learning the Precise Feature for Cluster Assignment
- Authors: Yanhai Gan, Xinghui Dong, Huiyu Zhou, Feng Gao, Junyu Dong
- Abstract summary: We propose a framework which integrates representation learning and clustering into a single pipeline for the first time.
The proposed framework exploits the powerful ability of recently developed generative models for learning intrinsic features.
Experimental results show that the performance of the proposed method is superior, or at least comparable to, the state-of-the-art methods.
- Score: 39.320210567860485
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Clustering is one of the fundamental tasks in computer vision and pattern
recognition. Recently, deep clustering methods (algorithms based on deep
learning) have attracted wide attention with their impressive performance. Most
of these algorithms combine deep unsupervised representation learning and
standard clustering together. However, the separation of representation
learning and clustering will lead to suboptimal solutions because the two-stage
strategy prevents representation learning from adapting to subsequent tasks
(e.g., clustering according to specific cues). To overcome this issue, efforts
have been made in the dynamic adaption of representation and cluster
assignment, whereas current state-of-the-art methods suffer from heuristically
constructed objectives with representation and cluster assignment alternatively
optimized. To further standardize the clustering problem, we audaciously
formulate the objective of clustering as finding a precise feature as the cue
for cluster assignment. Based on this, we propose a general-purpose deep
clustering framework which radically integrates representation learning and
clustering into a single pipeline for the first time. The proposed framework
exploits the powerful ability of recently developed generative models for
learning intrinsic features, and imposes an entropy minimization on the
distribution of the cluster assignment by a dedicated variational algorithm.
Experimental results show that the performance of the proposed method is
superior, or at least comparable to, the state-of-the-art methods on the
handwritten digit recognition, fashion recognition, face recognition and object
recognition benchmark datasets.
Related papers
- Image Clustering Algorithm Based on Self-Supervised Pretrained Models and Latent Feature Distribution Optimization [4.39139858370436]
This paper introduces an image clustering algorithm based on self-supervised pretrained models and latent feature distribution optimization.
Our approach outperforms the latest clustering algorithms and achieves state-of-the-art clustering results.
arXiv Detail & Related papers (2024-08-04T04:08:21Z) - GCC: Generative Calibration Clustering [55.44944397168619]
We propose a novel Generative Clustering (GCC) method to incorporate feature learning and augmentation into clustering procedure.
First, we develop a discrimirative feature alignment mechanism to discover intrinsic relationship across real and generated samples.
Second, we design a self-supervised metric learning to generate more reliable cluster assignment.
arXiv Detail & Related papers (2024-04-14T01:51:11Z) - Stable Cluster Discrimination for Deep Clustering [7.175082696240088]
Deep clustering can optimize representations of instances (i.e., representation learning) and explore the inherent data distribution.
The coupled objective implies a trivial solution that all instances collapse to the uniform features.
In this work, we first show that the prevalent discrimination task in supervised learning is unstable for one-stage clustering.
A novel stable cluster discrimination (SeCu) task is proposed and a new hardness-aware clustering criterion can be obtained accordingly.
arXiv Detail & Related papers (2023-11-24T06:43:26Z) - Reinforcement Graph Clustering with Unknown Cluster Number [91.4861135742095]
We propose a new deep graph clustering method termed Reinforcement Graph Clustering.
In our proposed method, cluster number determination and unsupervised representation learning are unified into a uniform framework.
In order to conduct feedback actions, the clustering-oriented reward function is proposed to enhance the cohesion of the same clusters and separate the different clusters.
arXiv Detail & Related papers (2023-08-13T18:12:28Z) - CEIL: A General Classification-Enhanced Iterative Learning Framework for
Text Clustering [16.08402937918212]
We propose a novel Classification-Enhanced Iterative Learning framework for short text clustering.
In each iteration, we first adopt a language model to retrieve the initial text representations.
After strict data filtering and aggregation processes, samples with clean category labels are retrieved, which serve as supervision information.
Finally, the updated language model with improved representation ability is used to enhance clustering in the next iteration.
arXiv Detail & Related papers (2023-04-20T14:04:31Z) - Unified Multi-View Orthonormal Non-Negative Graph Based Clustering
Framework [74.25493157757943]
We formulate a novel clustering model, which exploits the non-negative feature property and incorporates the multi-view information into a unified joint learning framework.
We also explore, for the first time, the multi-model non-negative graph-based approach to clustering data based on deep features.
arXiv Detail & Related papers (2022-11-03T08:18:27Z) - Rethinking Clustering-Based Pseudo-Labeling for Unsupervised
Meta-Learning [146.11600461034746]
Method for unsupervised meta-learning, CACTUs, is a clustering-based approach with pseudo-labeling.
This approach is model-agnostic and can be combined with supervised algorithms to learn from unlabeled data.
We prove that the core reason for this is lack of a clustering-friendly property in the embedding space.
arXiv Detail & Related papers (2022-09-27T19:04:36Z) - Representation Learning for Clustering via Building Consensus [3.7434090710577608]
We propose Consensus Clustering using Unsupervised Representation Learning (ConCURL)
ConCURL improves the clustering performance over state-of-the art methods on four out of five image datasets.
We extend the evaluation procedure for clustering to reflect the challenges in real world clustering tasks.
arXiv Detail & Related papers (2021-05-04T05:04:03Z) - Graph Contrastive Clustering [131.67881457114316]
We propose a novel graph contrastive learning framework, which is then applied to the clustering task and we come up with the Graph Constrastive Clustering(GCC) method.
Specifically, on the one hand, the graph Laplacian based contrastive loss is proposed to learn more discriminative and clustering-friendly features.
On the other hand, a novel graph-based contrastive learning strategy is proposed to learn more compact clustering assignments.
arXiv Detail & Related papers (2021-04-03T15:32:49Z) - Scalable Hierarchical Agglomerative Clustering [65.66407726145619]
Existing scalable hierarchical clustering methods sacrifice quality for speed.
We present a scalable, agglomerative method for hierarchical clustering that does not sacrifice quality and scales to billions of data points.
arXiv Detail & Related papers (2020-10-22T15:58:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.