Deep Clustering with a Constraint for Topological Invariance based on
Symmetric InfoNCE
- URL: http://arxiv.org/abs/2303.03036v1
- Date: Mon, 6 Mar 2023 11:05:21 GMT
- Title: Deep Clustering with a Constraint for Topological Invariance based on
Symmetric InfoNCE
- Authors: Yuhui Zhang, Yuichiro Wada, Hiroki Waida, Kaito Goto, Yusaku Hino,
Takafumi Kanamori
- Abstract summary: In this scenario, few existing state-of-the-art deep clustering methods can perform well for both non-complex topology and complex topology datasets.
We propose a constraint utilizing symmetric InfoNCE, which helps an objective of deep clustering method in the scenario train the model.
To confirm the effectiveness of the proposed constraint, we introduce a deep clustering method named MIST, which is a combination of an existing deep clustering method and our constraint.
- Score: 10.912082501425944
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We consider the scenario of deep clustering, in which the available prior
knowledge is limited. In this scenario, few existing state-of-the-art deep
clustering methods can perform well for both non-complex topology and complex
topology datasets. To address the problem, we propose a constraint utilizing
symmetric InfoNCE, which helps an objective of deep clustering method in the
scenario train the model so as to be efficient for not only non-complex
topology but also complex topology datasets. Additionally, we provide several
theoretical explanations of the reason why the constraint can enhances
performance of deep clustering methods. To confirm the effectiveness of the
proposed constraint, we introduce a deep clustering method named MIST, which is
a combination of an existing deep clustering method and our constraint. Our
numerical experiments via MIST demonstrate that the constraint is effective. In
addition, MIST outperforms other state-of-the-art deep clustering methods for
most of the commonly used ten benchmark datasets.
Related papers
- Robust and Automatic Data Clustering: Dirichlet Process meets
Median-of-Means [18.3248037914529]
We present an efficient and automatic clustering technique by integrating the principles of model-based and centroid-based methodologies.
Statistical guarantees on the upper bound of clustering error suggest the advantages of our proposed method over existing state-of-the-art clustering algorithms.
arXiv Detail & Related papers (2023-11-26T19:01:15Z) - DivClust: Controlling Diversity in Deep Clustering [47.85350249697335]
DivClust produces consensus clustering solutions that consistently outperform single-clustering baselines.
Our method effectively controls diversity across frameworks and datasets with very small additional computational cost.
arXiv Detail & Related papers (2023-04-03T14:45:43Z) - Deep Clustering: A Comprehensive Survey [53.387957674512585]
Clustering analysis plays an indispensable role in machine learning and data mining.
Deep clustering, which can learn clustering-friendly representations using deep neural networks, has been broadly applied in a wide range of clustering tasks.
Existing surveys for deep clustering mainly focus on the single-view fields and the network architectures, ignoring the complex application scenarios of clustering.
arXiv Detail & Related papers (2022-10-09T02:31:32Z) - A One-shot Framework for Distributed Clustered Learning in Heterogeneous
Environments [54.172993875654015]
The paper proposes a family of communication efficient methods for distributed learning in heterogeneous environments.
One-shot approach, based on local computations at the users and a clustering based aggregation step at the server is shown to provide strong learning guarantees.
For strongly convex problems it is shown that, as long as the number of data points per user is above a threshold, the proposed approach achieves order-optimal mean-squared error rates in terms of the sample size.
arXiv Detail & Related papers (2022-09-22T09:04:10Z) - Deep Conditional Gaussian Mixture Model for Constrained Clustering [7.070883800886882]
Constrained clustering can leverage prior information on a growing amount of only partially labeled data.
We propose a novel framework for constrained clustering that is intuitive, interpretable, and can be trained efficiently in the framework of gradient variational inference.
arXiv Detail & Related papers (2021-06-11T13:38:09Z) - A Deep Learning Object Detection Method for an Efficient Clusters
Initialization [6.365889364810239]
Clustering has been used in numerous applications such as banking customers profiling, document retrieval, image segmentation, and e-commerce recommendation engines.
Existing clustering techniques present significant limitations, from which is the dependability of their stability on the initialization parameters.
This paper proposes a solution that can provide near-optimal clustering parameters with low computational and resources overhead.
arXiv Detail & Related papers (2021-04-28T08:34:25Z) - A Framework for Deep Constrained Clustering [19.07636653413663]
Constrained clustering formulations exist for popular algorithms such as k-means, mixture models, and spectral clustering but have several limitations.
Here we explore a deep learning framework for constrained clustering and in particular explore how it can extend the field of constrained clustering.
We show that our framework can not only handle standard together/apart constraints (without the well documented negative effects reported earlier) generated from labeled side information.
We propose an efficient training paradigm that is generally applicable to these four types of constraints.
arXiv Detail & Related papers (2021-01-07T22:49:06Z) - Overcomplete Deep Subspace Clustering Networks [80.16644725886968]
Experimental results on four benchmark datasets show the effectiveness of the proposed method over DSC and other clustering methods in terms of clustering error.
Our method is also not as dependent as DSC is on where pre-training should be stopped to get the best performance and is also more robust to noise.
arXiv Detail & Related papers (2020-11-16T22:07:18Z) - Scalable Hierarchical Agglomerative Clustering [65.66407726145619]
Existing scalable hierarchical clustering methods sacrifice quality for speed.
We present a scalable, agglomerative method for hierarchical clustering that does not sacrifice quality and scales to billions of data points.
arXiv Detail & Related papers (2020-10-22T15:58:35Z) - Unsupervised Multi-view Clustering by Squeezing Hybrid Knowledge from
Cross View and Each View [68.88732535086338]
This paper proposes a new multi-view clustering method, low-rank subspace multi-view clustering based on adaptive graph regularization.
Experimental results for five widely used multi-view benchmarks show that our proposed algorithm surpasses other state-of-the-art methods by a clear margin.
arXiv Detail & Related papers (2020-08-23T08:25:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.