Deep Robust Clustering by Contrastive Learning
- URL: http://arxiv.org/abs/2008.03030v2
- Date: Thu, 27 Aug 2020 05:42:48 GMT
- Title: Deep Robust Clustering by Contrastive Learning
- Authors: Huasong Zhong, Chong Chen, Zhongming Jin, Xian-Sheng Hua
- Abstract summary: We propose Deep Robust Clustering (DRC) to learn clustering with unlabelled data.
DRC looks into deep clustering from two perspectives of both semantic clustering assignment and representation feature.
Experiments on six widely-adopted deep clustering benchmarks demonstrate the superiority of DRC in both stability and accuracy.
- Score: 31.161207608881472
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recently, many unsupervised deep learning methods have been proposed to learn
clustering with unlabelled data. By introducing data augmentation, most of the
latest methods look into deep clustering from the perspective that the original
image and its transformation should share similar semantic clustering
assignment. However, the representation features could be quite different even
they are assigned to the same cluster since softmax function is only sensitive
to the maximum value. This may result in high intra-class diversities in the
representation feature space, which will lead to unstable local optimal and
thus harm the clustering performance. To address this drawback, we proposed
Deep Robust Clustering (DRC). Different from existing methods, DRC looks into
deep clustering from two perspectives of both semantic clustering assignment
and representation feature, which can increase inter-class diversities and
decrease intra-class diversities simultaneously. Furthermore, we summarized a
general framework that can turn any maximizing mutual information into
minimizing contrastive loss by investigating the internal relationship between
mutual information and contrastive learning. And we successfully applied it in
DRC to learn invariant features and robust clusters. Extensive experiments on
six widely-adopted deep clustering benchmarks demonstrate the superiority of
DRC in both stability and accuracy. e.g., attaining 71.6% mean accuracy on
CIFAR-10, which is 7.1% higher than state-of-the-art results.
Related papers
- Self-Supervised Graph Embedding Clustering [70.36328717683297]
K-means one-step dimensionality reduction clustering method has made some progress in addressing the curse of dimensionality in clustering tasks.
We propose a unified framework that integrates manifold learning with K-means, resulting in the self-supervised graph embedding framework.
arXiv Detail & Related papers (2024-09-24T08:59:51Z) - New Intent Discovery with Attracting and Dispersing Prototype [32.607743914487614]
New Intent Discovery (NID) aims to recognize known and infer new intent categories with the help of limited labeled and large-scale unlabeled data.
We propose a Robust and Adaptive Prototypical learning framework for globally distinct decision boundaries for both known and new intent categories.
arXiv Detail & Related papers (2024-03-25T16:31:55Z) - Stable Cluster Discrimination for Deep Clustering [7.175082696240088]
Deep clustering can optimize representations of instances (i.e., representation learning) and explore the inherent data distribution.
The coupled objective implies a trivial solution that all instances collapse to the uniform features.
In this work, we first show that the prevalent discrimination task in supervised learning is unstable for one-stage clustering.
A novel stable cluster discrimination (SeCu) task is proposed and a new hardness-aware clustering criterion can be obtained accordingly.
arXiv Detail & Related papers (2023-11-24T06:43:26Z) - Rethinking k-means from manifold learning perspective [122.38667613245151]
We present a new clustering algorithm which directly detects clusters of data without mean estimation.
Specifically, we construct distance matrix between data points by Butterworth filter.
To well exploit the complementary information embedded in different views, we leverage the tensor Schatten p-norm regularization.
arXiv Detail & Related papers (2023-05-12T03:01:41Z) - Deep Clustering of Tabular Data by Weighted Gaussian Distribution Learning [0.0]
This paper develops one of the first deep clustering methods for tabular data: Gaussian Cluster Embedding in Autoencoder Latent Space (G-CEALS)
The G-CEALS method presents average rank orderings of 2.9(1.7) and 2.8(1.7) based on clustering accuracy and adjusted Rand index (ARI) scores on sixteen data sets, respectively, and outperforms nine state-of-the-art clustering methods.
arXiv Detail & Related papers (2023-01-02T18:45:53Z) - Semi-supervised Domain Adaptive Structure Learning [72.01544419893628]
Semi-supervised domain adaptation (SSDA) is a challenging problem requiring methods to overcome both 1) overfitting towards poorly annotated data and 2) distribution shift across domains.
We introduce an adaptive structure learning method to regularize the cooperation of SSL and DA.
arXiv Detail & Related papers (2021-12-12T06:11:16Z) - Cluster Analysis with Deep Embeddings and Contrastive Learning [0.0]
This work proposes a novel framework for performing image clustering from deep embeddings.
Our approach jointly learns representations and predicts cluster centers in an end-to-end manner.
Our framework performs on par with widely accepted clustering methods and outperforms the state-of-the-art contrastive learning method on the CIFAR-10 dataset.
arXiv Detail & Related papers (2021-09-26T22:18:15Z) - Learning Statistical Representation with Joint Deep Embedded Clustering [2.1267423178232407]
StatDEC is an unsupervised framework for joint statistical representation learning and clustering.
Our experiments show that using these representations, one can considerably improve results on imbalanced image clustering across a variety of image datasets.
arXiv Detail & Related papers (2021-09-11T09:26:52Z) - You Never Cluster Alone [150.94921340034688]
We extend the mainstream contrastive learning paradigm to a cluster-level scheme, where all the data subjected to the same cluster contribute to a unified representation.
We define a set of categorical variables as clustering assignment confidence, which links the instance-level learning track with the cluster-level one.
By reparametrizing the assignment variables, TCC is trained end-to-end, requiring no alternating steps.
arXiv Detail & Related papers (2021-06-03T14:59:59Z) - Graph Contrastive Clustering [131.67881457114316]
We propose a novel graph contrastive learning framework, which is then applied to the clustering task and we come up with the Graph Constrastive Clustering(GCC) method.
Specifically, on the one hand, the graph Laplacian based contrastive loss is proposed to learn more discriminative and clustering-friendly features.
On the other hand, a novel graph-based contrastive learning strategy is proposed to learn more compact clustering assignments.
arXiv Detail & Related papers (2021-04-03T15:32:49Z) - Contrastive Clustering [57.71729650297379]
We propose Contrastive Clustering (CC) which explicitly performs the instance- and cluster-level contrastive learning.
In particular, CC achieves an NMI of 0.705 (0.431) on the CIFAR-10 (CIFAR-100) dataset, which is an up to 19% (39%) performance improvement compared with the best baseline.
arXiv Detail & Related papers (2020-09-21T08:54:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.