Oracle-guided Contrastive Clustering
- URL: http://arxiv.org/abs/2211.00409v1
- Date: Tue, 1 Nov 2022 12:05:12 GMT
- Title: Oracle-guided Contrastive Clustering
- Authors: Mengdie Wang, Liyuan Shang, Suyun Zhao, Yiming Wang, Hong Chen,
Cuiping Li and Xizhao Wang
- Abstract summary: Oracle-guided Contrastive Clustering(OCC) is proposed to cluster by interactively making pairwise same-cluster" queries to oracles with distinctive demands.
To the best of our knowledge, it is the first deep framework to perform personalized clustering.
- Score: 28.066047266687058
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Deep clustering aims to learn a clustering representation through deep
architectures. Most of the existing methods usually conduct clustering with the
unique goal of maximizing clustering performance, that ignores the personalized
demand of clustering tasks.% and results in unguided clustering solutions.
However, in real scenarios, oracles may tend to cluster unlabeled data by
exploiting distinct criteria, such as distinct semantics (background, color,
object, etc.), and then put forward personalized clustering tasks. To achieve
task-aware clustering results, in this study, Oracle-guided Contrastive
Clustering(OCC) is then proposed to cluster by interactively making pairwise
``same-cluster" queries to oracles with distinctive demands. Specifically,
inspired by active learning, some informative instance pairs are queried, and
evaluated by oracles whether the pairs are in the same cluster according to
their desired orientation. And then these queried same-cluster pairs extend the
set of positive instance pairs for contrastive learning, guiding OCC to extract
orientation-aware feature representation. Accordingly, the query results,
guided by oracles with distinctive demands, may drive the OCC's clustering
results in a desired orientation. Theoretically, the clustering risk in an
active learning manner is given with a tighter upper bound, that guarantees
active queries to oracles do mitigate the clustering risk. Experimentally,
extensive results verify that OCC can cluster accurately along the specific
orientation and it substantially outperforms the SOTA clustering methods as
well. To the best of our knowledge, it is the first deep framework to perform
personalized clustering.
Related papers
- Towards Explainable Clustering: A Constrained Declarative based Approach [0.294944680995069]
We aim at finding a clustering that has high quality in terms of classic clustering criteria and that is explainable.
A good global explanation of a clustering should give the characteristics of each cluster taking into account their abilities to describe its objects.
We propose a novel interpretable constrained method called ECS for declarative computation with Explainabilty-driven Selection.
arXiv Detail & Related papers (2024-03-26T21:00:06Z) - Deep Embedding Clustering Driven by Sample Stability [16.53706617383543]
We propose a deep embedding clustering algorithm driven by sample stability (DECS)
Specifically, we start by constructing the initial feature space with an autoencoder and then learn the cluster-oriented embedding feature constrained by sample stability.
The experimental results on five datasets illustrate that the proposed method achieves superior performance compared to state-of-the-art clustering approaches.
arXiv Detail & Related papers (2024-01-29T09:19:49Z) - Stable Cluster Discrimination for Deep Clustering [7.175082696240088]
Deep clustering can optimize representations of instances (i.e., representation learning) and explore the inherent data distribution.
The coupled objective implies a trivial solution that all instances collapse to the uniform features.
In this work, we first show that the prevalent discrimination task in supervised learning is unstable for one-stage clustering.
A novel stable cluster discrimination (SeCu) task is proposed and a new hardness-aware clustering criterion can be obtained accordingly.
arXiv Detail & Related papers (2023-11-24T06:43:26Z) - Reinforcement Graph Clustering with Unknown Cluster Number [91.4861135742095]
We propose a new deep graph clustering method termed Reinforcement Graph Clustering.
In our proposed method, cluster number determination and unsupervised representation learning are unified into a uniform framework.
In order to conduct feedback actions, the clustering-oriented reward function is proposed to enhance the cohesion of the same clusters and separate the different clusters.
arXiv Detail & Related papers (2023-08-13T18:12:28Z) - Instance-Optimal Cluster Recovery in the Labeled Stochastic Block Model [79.46465138631592]
We devise an efficient algorithm that recovers clusters using the observed labels.
We present Instance-Adaptive Clustering (IAC), the first algorithm whose performance matches these lower bounds both in expectation and with high probability.
arXiv Detail & Related papers (2023-06-18T08:46:06Z) - Self-supervised Contrastive Attributed Graph Clustering [110.52694943592974]
We propose a novel attributed graph clustering network, namely Self-supervised Contrastive Attributed Graph Clustering (SCAGC)
In SCAGC, by leveraging inaccurate clustering labels, a self-supervised contrastive loss, are designed for node representation learning.
For the OOS nodes, SCAGC can directly calculate their clustering labels.
arXiv Detail & Related papers (2021-10-15T03:25:28Z) - Learning to Cluster via Same-Cluster Queries [26.284461833343403]
We study the problem of learning to cluster data points using an oracle which can answer same-cluster queries.
We propose two algorithms with provable theoretical guarantees and verify their effectiveness via an extensive set of experiments on both synthetic and real-world data.
arXiv Detail & Related papers (2021-08-17T00:37:11Z) - You Never Cluster Alone [150.94921340034688]
We extend the mainstream contrastive learning paradigm to a cluster-level scheme, where all the data subjected to the same cluster contribute to a unified representation.
We define a set of categorical variables as clustering assignment confidence, which links the instance-level learning track with the cluster-level one.
By reparametrizing the assignment variables, TCC is trained end-to-end, requiring no alternating steps.
arXiv Detail & Related papers (2021-06-03T14:59:59Z) - Scalable Hierarchical Agglomerative Clustering [65.66407726145619]
Existing scalable hierarchical clustering methods sacrifice quality for speed.
We present a scalable, agglomerative method for hierarchical clustering that does not sacrifice quality and scales to billions of data points.
arXiv Detail & Related papers (2020-10-22T15:58:35Z) - Contrastive Clustering [57.71729650297379]
We propose Contrastive Clustering (CC) which explicitly performs the instance- and cluster-level contrastive learning.
In particular, CC achieves an NMI of 0.705 (0.431) on the CIFAR-10 (CIFAR-100) dataset, which is an up to 19% (39%) performance improvement compared with the best baseline.
arXiv Detail & Related papers (2020-09-21T08:54:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.