Spatial Transformer K-Means
- URL: http://arxiv.org/abs/2202.07829v1
- Date: Wed, 16 Feb 2022 02:25:46 GMT
- Title: Spatial Transformer K-Means
- Authors: Romain Cosentino, Randall Balestriero, Yanis Bahroun, Anirvan
Sengupta, Richard Baraniuk, Behnaam Aazhang
- Abstract summary: Intricate data embeddings have been designed to push $K$-means performances.
We propose preserving the intrinsic data space and augment K-means with a similarity measure invariant to non-rigid transformations.
- Score: 16.775789494555017
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: K-means defines one of the most employed centroid-based clustering algorithms
with performances tied to the data's embedding. Intricate data embeddings have
been designed to push $K$-means performances at the cost of reduced theoretical
guarantees and interpretability of the results. Instead, we propose preserving
the intrinsic data space and augment K-means with a similarity measure
invariant to non-rigid transformations. This enables (i) the reduction of
intrinsic nuisances associated with the data, reducing the complexity of the
clustering task and increasing performances and producing state-of-the-art
results, (ii) clustering in the input space of the data, leading to a fully
interpretable clustering algorithm, and (iii) the benefit of convergence
guarantees.
Related papers
- Adaptively Robust and Sparse K-means Clustering [5.535948428518607]
This paper proposes adaptively robust and sparse K-means clustering (ARSK) to address these practical limitations of the standard K-means algorithm.
We introduce a redundant error component for each observation for robustness, and this additional parameter is penalized using a group sparse penalty.
To accommodate the impact of high-dimensional noisy variables, the objective function is modified by incorporating weights and implementing a penalty to control the sparsity of the weight vector.
arXiv Detail & Related papers (2024-07-09T15:20:41Z) - Fuzzy K-Means Clustering without Cluster Centroids [79.19713746387337]
Fuzzy K-Means clustering is a critical computation technique in unsupervised data analysis.
This paper proposes a novel Fuzzy K-Means clustering algorithm that entirely eliminates the reliance on cluster centroids.
arXiv Detail & Related papers (2024-04-07T12:25:03Z) - Federated K-means Clustering [0.0]
Federated learning is a technique that enables the use of distributed datasets for machine learning purposes without requiring data to be pooled.
This work introduces an algorithm which implements K-means clustering in a federated manner.
arXiv Detail & Related papers (2023-10-02T13:32:00Z) - Rethinking k-means from manifold learning perspective [122.38667613245151]
We present a new clustering algorithm which directly detects clusters of data without mean estimation.
Specifically, we construct distance matrix between data points by Butterworth filter.
To well exploit the complementary information embedded in different views, we leverage the tensor Schatten p-norm regularization.
arXiv Detail & Related papers (2023-05-12T03:01:41Z) - Differentially Private Federated Clustering over Non-IID Data [59.611244450530315]
clustering clusters (FedC) problem aims to accurately partition unlabeled data samples distributed over massive clients into finite clients under the orchestration of a server.
We propose a novel FedC algorithm using differential privacy convergence technique, referred to as DP-Fed, in which partial participation and multiple clients are also considered.
Various attributes of the proposed DP-Fed are obtained through theoretical analyses of privacy protection, especially for the case of non-identically and independently distributed (non-i.i.d.) data.
arXiv Detail & Related papers (2023-01-03T05:38:43Z) - Fast and Interpretable Consensus Clustering via Minipatch Learning [0.0]
We develop IMPACC: Interpretable MiniPatch Adaptive Consensus Clustering.
We develop adaptive sampling schemes for observations, which result in both improved reliability and computational savings.
Results show that our approach yields more accurate and interpretable cluster solutions.
arXiv Detail & Related papers (2021-10-05T22:39:28Z) - Automated Clustering of High-dimensional Data with a Feature Weighted
Mean Shift Algorithm [16.0817847880416]
Mean shift is a simple interactive procedure that shifts data points towards the mode which denotes the highest density of data points in the region.
We propose a simple yet elegant feature-weighted variant of mean shift to efficiently learn the feature importance.
The resulting algorithm not only outperforms the conventional mean shift clustering procedure but also preserves its computational simplicity.
arXiv Detail & Related papers (2020-12-20T14:00:40Z) - Effective Data-aware Covariance Estimator from Compressed Data [63.16042585506435]
We propose a data-aware weighted sampling based covariance matrix estimator, namely DACE, which can provide an unbiased covariance matrix estimation.
We conduct extensive experiments on both synthetic and real-world datasets to demonstrate the superior performance of our DACE.
arXiv Detail & Related papers (2020-10-10T10:10:28Z) - Decorrelated Clustering with Data Selection Bias [55.91842043124102]
We propose a novel Decorrelation regularized K-Means algorithm (DCKM) for clustering with data selection bias.
Our DCKM algorithm achieves significant performance gains, indicating the necessity of removing unexpected feature correlations induced by selection bias.
arXiv Detail & Related papers (2020-06-29T08:55:50Z) - New advances in enumerative biclustering algorithms with online
partitioning [80.22629846165306]
This paper further extends RIn-Close_CVC, a biclustering algorithm capable of performing an efficient, complete, correct and non-redundant enumeration of maximal biclusters with constant values on columns in numerical datasets.
The improved algorithm is called RIn-Close_CVC3, keeps those attractive properties of RIn-Close_CVC, and is characterized by: a drastic reduction in memory usage; a consistent gain in runtime.
arXiv Detail & Related papers (2020-03-07T14:54:26Z) - Simple and Scalable Sparse k-means Clustering via Feature Ranking [14.839931533868176]
We propose a novel framework for sparse k-means clustering that is intuitive, simple to implement, and competitive with state-of-the-art algorithms.
Our core method readily generalizes to several task-specific algorithms such as clustering on subsets of attributes and in partially observed data settings.
arXiv Detail & Related papers (2020-02-20T02:41:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.