Parametric entropy based Cluster Centriod Initialization for k-means
clustering of various Image datasets
- URL: http://arxiv.org/abs/2308.07705v1
- Date: Tue, 15 Aug 2023 11:28:02 GMT
- Title: Parametric entropy based Cluster Centriod Initialization for k-means
clustering of various Image datasets
- Authors: Faheem Hussayn and Shahid M Shah
- Abstract summary: k-means is one of the most employed yet simple algorithms for cluster analysis.
In this paper, we conduct an analysis of the performance of k-means on image data by employing parametric entropies.
We propose the best fitting entropy measures for general image datasets.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: One of the most employed yet simple algorithm for cluster analysis is the
k-means algorithm. k-means has successfully witnessed its use in artificial
intelligence, market segmentation, fraud detection, data mining, psychology,
etc., only to name a few. The k-means algorithm, however, does not always yield
the best quality results. Its performance heavily depends upon the number of
clusters supplied and the proper initialization of the cluster centroids or
seeds. In this paper, we conduct an analysis of the performance of k-means on
image data by employing parametric entropies in an entropy based centroid
initialization method and propose the best fitting entropy measures for general
image datasets. We use several entropies like Taneja entropy, Kapur entropy,
Aczel Daroczy entropy, Sharma Mittal entropy. We observe that for different
datasets, different entropies provide better results than the conventional
methods. We have applied our proposed algorithm on these datasets: Satellite,
Toys, Fruits, Cars, Brain MRI, Covid X-Ray.
Related papers
- Automated regime detection in multidimensional time series data using
sliced Wasserstein k-means clustering [0.0]
We study the behaviour of the Wasserstein k-means clustering algorithm applied to time series data.
We extend the technique to multidimensional time series data by approximating the multidimensional Wasserstein distance as a sliced Wasserstein distance.
We show that the sWk-means method is effective in identifying distinct market regimes in real multidimensional financial time series.
arXiv Detail & Related papers (2023-10-02T15:37:56Z) - Rethinking k-means from manifold learning perspective [122.38667613245151]
We present a new clustering algorithm which directly detects clusters of data without mean estimation.
Specifically, we construct distance matrix between data points by Butterworth filter.
To well exploit the complementary information embedded in different views, we leverage the tensor Schatten p-norm regularization.
arXiv Detail & Related papers (2023-05-12T03:01:41Z) - An enhanced method of initial cluster center selection for K-means
algorithm [0.0]
We propose a novel approach to improve initial cluster selection for K-means algorithm.
The Convex Hull algorithm facilitates the computing of the first two centroids and the remaining ones are selected according to the distance from previously selected centers.
We obtained only 7.33%, 7.90%, and 0% clustering error in Iris, Letter, and Ruspini data respectively.
arXiv Detail & Related papers (2022-10-18T00:58:50Z) - How to Use K-means for Big Data Clustering? [2.1165011830664677]
K-means is the simplest and most widely used algorithm under the Euclidean Minimum Sum-of-Squares Clustering (MSSC) model.
We propose a new parallel scheme of using K-means and K-means++ algorithms for big data clustering.
arXiv Detail & Related papers (2022-04-14T08:18:01Z) - Differentially-Private Clustering of Easy Instances [67.04951703461657]
In differentially private clustering, the goal is to identify $k$ cluster centers without disclosing information on individual data points.
We provide implementable differentially private clustering algorithms that provide utility when the data is "easy"
We propose a framework that allows us to apply non-private clustering algorithms to the easy instances and privately combine the results.
arXiv Detail & Related papers (2021-12-29T08:13:56Z) - Estimating leverage scores via rank revealing methods and randomization [50.591267188664666]
We study algorithms for estimating the statistical leverage scores of rectangular dense or sparse matrices of arbitrary rank.
Our approach is based on combining rank revealing methods with compositions of dense and sparse randomized dimensionality reduction transforms.
arXiv Detail & Related papers (2021-05-23T19:21:55Z) - Finding Geometric Models by Clustering in the Consensus Space [61.65661010039768]
We propose a new algorithm for finding an unknown number of geometric models, e.g., homographies.
We present a number of applications where the use of multiple geometric models improves accuracy.
These include pose estimation from multiple generalized homographies; trajectory estimation of fast-moving objects.
arXiv Detail & Related papers (2021-03-25T14:35:07Z) - Determinantal consensus clustering [77.34726150561087]
We propose the use of determinantal point processes or DPP for the random restart of clustering algorithms.
DPPs favor diversity of the center points within subsets.
We show through simulations that, contrary to DPP, this technique fails both to ensure diversity, and to obtain a good coverage of all data facets.
arXiv Detail & Related papers (2021-02-07T23:48:24Z) - Clustering of Big Data with Mixed Features [3.3504365823045044]
We develop a new clustering algorithm for large data of mixed type.
The algorithm is capable of detecting outliers and clusters of relatively lower density values.
We present experimental results to verify that our algorithm works well in practice.
arXiv Detail & Related papers (2020-11-11T19:54:38Z) - Differentially Private Clustering: Tight Approximation Ratios [57.89473217052714]
We give efficient differentially private algorithms for basic clustering problems.
Our results imply an improved algorithm for the Sample and Aggregate privacy framework.
One of the tools used in our 1-Cluster algorithm can be employed to get a faster quantum algorithm for ClosestPair in a moderate number of dimensions.
arXiv Detail & Related papers (2020-08-18T16:22:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.