A Centroid Auto-Fused Hierarchical Fuzzy c-Means Clustering
- URL: http://arxiv.org/abs/2004.12756v1
- Date: Mon, 27 Apr 2020 12:59:22 GMT
- Title: A Centroid Auto-Fused Hierarchical Fuzzy c-Means Clustering
- Authors: Yunxia Lin, Songcan Chen
- Abstract summary: Centroid Auto-Fused Hierarchical Fuzzy c-means method (CAF-HFCM)
We present a Centroid Auto-Fused Hierarchical Fuzzy c-means method (CAF-HFCM) whose optimization procedure can automatically agglomerate to form a cluster hierarchy.
Our proposed CAF-HFCM method is able to be straightforwardly extended to various variants of FCM.
- Score: 30.709797128259236
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Like k-means and Gaussian Mixture Model (GMM), fuzzy c-means (FCM) with soft
partition has also become a popular clustering algorithm and still is
extensively studied. However, these algorithms and their variants still suffer
from some difficulties such as determination of the optimal number of clusters
which is a key factor for clustering quality. A common approach for overcoming
this difficulty is to use the trial-and-validation strategy, i.e., traversing
every integer from large number like $\sqrt{n}$ to 2 until finding the optimal
number corresponding to the peak value of some cluster validity index. But it
is scarcely possible to naturally construct an adaptively agglomerative
hierarchical cluster structure as using the trial-and-validation strategy. Even
possible, existing different validity indices also lead to different number of
clusters. To effectively mitigate the problems while motivated by convex
clustering, in this paper we present a Centroid Auto-Fused Hierarchical Fuzzy
c-means method (CAF-HFCM) whose optimization procedure can automatically
agglomerate to form a cluster hierarchy, more importantly, yielding an optimal
number of clusters without resorting to any validity index. Although a
recently-proposed robust-learning fuzzy c-means (RL-FCM) can also automatically
obtain the best number of clusters without the help of any validity index,
so-involved 3 hyper-parameters need to adjust expensively, conversely, our
CAF-HFCM involves just 1 hyper-parameter which makes the corresponding
adjustment is relatively easier and more operational. Further, as an additional
benefit from our optimization objective, the CAF-HFCM effectively reduces the
sensitivity to the initialization of clustering performance. Moreover, our
proposed CAF-HFCM method is able to be straightforwardly extended to various
variants of FCM.
Related papers
- Self-Supervised Graph Embedding Clustering [70.36328717683297]
K-means one-step dimensionality reduction clustering method has made some progress in addressing the curse of dimensionality in clustering tasks.
We propose a unified framework that integrates manifold learning with K-means, resulting in the self-supervised graph embedding framework.
arXiv Detail & Related papers (2024-09-24T08:59:51Z) - MOKD: Cross-domain Finetuning for Few-shot Classification via Maximizing Optimized Kernel Dependence [97.93517982908007]
In cross-domain few-shot classification, NCC aims to learn representations to construct a metric space where few-shot classification can be performed.
In this paper, we find that there exist high similarities between NCC-learned representations of two samples from different classes.
We propose a bi-level optimization framework, emphmaximizing optimized kernel dependence (MOKD) to learn a set of class-specific representations that match the cluster structures indicated by labeled data.
arXiv Detail & Related papers (2024-05-29T05:59:52Z) - Fuzzy K-Means Clustering without Cluster Centroids [21.256564324236333]
Fuzzy K-Means clustering is a critical technique in unsupervised data analysis.
This paper proposes a novel Fuzzy textitK-Means clustering algorithm that entirely eliminates the reliance on cluster centroids.
arXiv Detail & Related papers (2024-04-07T12:25:03Z) - Instance-Optimal Cluster Recovery in the Labeled Stochastic Block Model [79.46465138631592]
We devise an efficient algorithm that recovers clusters using the observed labels.
We present Instance-Adaptive Clustering (IAC), the first algorithm whose performance matches these lower bounds both in expectation and with high probability.
arXiv Detail & Related papers (2023-06-18T08:46:06Z) - Accelerated Fuzzy C-Means Clustering Based on New Affinity Filtering and
Membership Scaling [74.85538972921917]
Fuzzy C-Means (FCM) is a widely used clustering method.
FCM has low efficiency in the mid-to-late stage of the clustering process.
FCM based on new affinity filtering and membership scaling (AMFCM) is proposed to accelerate the whole convergence process.
arXiv Detail & Related papers (2023-02-14T14:20:31Z) - Hybridization of K-means with improved firefly algorithm for automatic
clustering in high dimension [0.0]
We have implemented the Silhouette and Elbow methods with PCA to find an optimal number of clusters.
In the Firefly algorithm, the entire population is automatically subdivided into sub-populations that decrease the convergence rate speed and trapping to local minima.
Our study proposed an enhanced firefly, i.e., a hybridized K-means with an ODFA model for automatic clustering.
arXiv Detail & Related papers (2023-02-09T18:43:10Z) - An Improved Algorithm for Clustered Federated Learning [29.166363192740768]
This paper addresses the dichotomy between heterogeneous models and simultaneous training in Federated Learning (FL)
We define a new clustering model for FL based on the (optimal) local models of the users.
textttSR-FCA uses a robust learning algorithm within each cluster to exploit simultaneous training and to correct clustering errors.
arXiv Detail & Related papers (2022-10-20T19:14:36Z) - Late Fusion Multi-view Clustering via Global and Local Alignment
Maximization [61.89218392703043]
Multi-view clustering (MVC) optimally integrates complementary information from different views to improve clustering performance.
Most of existing approaches directly fuse multiple pre-specified similarities to learn an optimal similarity matrix for clustering.
We propose late fusion MVC via alignment to address these issues.
arXiv Detail & Related papers (2022-08-02T01:49:31Z) - Scalable Hierarchical Agglomerative Clustering [65.66407726145619]
Existing scalable hierarchical clustering methods sacrifice quality for speed.
We present a scalable, agglomerative method for hierarchical clustering that does not sacrifice quality and scales to billions of data points.
arXiv Detail & Related papers (2020-10-22T15:58:35Z) - A New Validity Index for Fuzzy-Possibilistic C-Means Clustering [6.174448419090291]
Fuzzy-Possibilistic (FP) index works well in the presence of clusters that vary in shape and density.
FPCM requires a priori selection of the degree of fuzziness and the degree of typicality.
arXiv Detail & Related papers (2020-05-19T01:48:13Z) - Supervised Enhanced Soft Subspace Clustering (SESSC) for TSK Fuzzy
Classifiers [25.32478253796209]
Fuzzy c-means based clustering algorithms are frequently used for Takagi-Sugeno-Kang (TSK) fuzzy classifier parameter estimation.
This paper proposes a supervised enhanced soft subspace clustering (SESSC) algorithm, which considers simultaneously the within-cluster compactness, between-cluster separation, and label information in clustering.
arXiv Detail & Related papers (2020-02-27T19:39:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.