Kernel Correlation-Dissimilarity for Multiple Kernel k-Means Clustering
- URL: http://arxiv.org/abs/2403.03448v1
- Date: Wed, 6 Mar 2024 04:24:43 GMT
- Title: Kernel Correlation-Dissimilarity for Multiple Kernel k-Means Clustering
- Authors: Rina Su, Yu Guo, Caiying Wu, Qiyu Jin, Tieyong Zeng
- Abstract summary: Current methods enhance information diversity and reduce redundancy by exploiting interdependencies among multiple kernels based on correlations or dissimilarities.
We introduce a novel method that systematically integrates both kernel correlation and dissimilarity.
By emphasizing the coherence between kernel correlation and dissimilarity, our method offers a more objective and transparent strategy for extracting non-linear information.
- Score: 21.685153346752124
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The main objective of the Multiple Kernel k-Means (MKKM) algorithm is to
extract non-linear information and achieve optimal clustering by optimizing
base kernel matrices. Current methods enhance information diversity and reduce
redundancy by exploiting interdependencies among multiple kernels based on
correlations or dissimilarities. Nevertheless, relying solely on a single
metric, such as correlation or dissimilarity, to define kernel relationships
introduces bias and incomplete characterization. Consequently, this limitation
hinders efficient information extraction, ultimately compromising clustering
performance. To tackle this challenge, we introduce a novel method that
systematically integrates both kernel correlation and dissimilarity. Our
approach comprehensively captures kernel relationships, facilitating more
efficient classification information extraction and improving clustering
performance. By emphasizing the coherence between kernel correlation and
dissimilarity, our method offers a more objective and transparent strategy for
extracting non-linear information and significantly improving clustering
precision, supported by theoretical rationale. We assess the performance of our
algorithm on 13 challenging benchmark datasets, demonstrating its superiority
over contemporary state-of-the-art MKKM techniques.
Related papers
- Boosting the Performance of Decentralized Federated Learning via Catalyst Acceleration [66.43954501171292]
We introduce Catalyst Acceleration and propose an acceleration Decentralized Federated Learning algorithm called DFedCata.
DFedCata consists of two main components: the Moreau envelope function, which addresses parameter inconsistencies, and Nesterov's extrapolation step, which accelerates the aggregation phase.
Empirically, we demonstrate the advantages of the proposed algorithm in both convergence speed and generalization performance on CIFAR10/100 with various non-iid data distributions.
arXiv Detail & Related papers (2024-10-09T06:17:16Z) - Self-Supervised Graph Embedding Clustering [70.36328717683297]
K-means one-step dimensionality reduction clustering method has made some progress in addressing the curse of dimensionality in clustering tasks.
We propose a unified framework that integrates manifold learning with K-means, resulting in the self-supervised graph embedding framework.
arXiv Detail & Related papers (2024-09-24T08:59:51Z) - Fast Asymmetric Factorization for Large Scale Multiple Kernel Clustering [5.21777096853979]
Multiple Kernel Clustering (MKC) has emerged as a solution, allowing the fusion of information from multiple base kernels for clustering.
We propose Efficient Multiple Kernel Concept Factorization (EMKCF), which constructs a new sparse kernel matrix inspired by local regression to achieve memory efficiency.
arXiv Detail & Related papers (2024-05-26T06:29:12Z) - Bridging Algorithmic Information Theory and Machine Learning: A New Approach to Kernel Learning [12.848057726330714]
We adopt an AIT perspective on the problem of learning kernels from data, through the method of Sparse Kernel Flows.
This approach opens the door to reformulating algorithms in machine learning using tools from AIT.
arXiv Detail & Related papers (2023-11-21T14:18:28Z) - Linear Self-Attention Approximation via Trainable Feedforward Kernel [77.34726150561087]
In pursuit of faster computation, Efficient Transformers demonstrate an impressive variety of approaches.
We aim to expand the idea of trainable kernel methods to approximate the self-attention mechanism of the Transformer architecture.
arXiv Detail & Related papers (2022-11-08T08:14:11Z) - Rethinking Clustering-Based Pseudo-Labeling for Unsupervised
Meta-Learning [146.11600461034746]
Method for unsupervised meta-learning, CACTUs, is a clustering-based approach with pseudo-labeling.
This approach is model-agnostic and can be combined with supervised algorithms to learn from unlabeled data.
We prove that the core reason for this is lack of a clustering-friendly property in the embedding space.
arXiv Detail & Related papers (2022-09-27T19:04:36Z) - Recovery Guarantees for Kernel-based Clustering under Non-parametric
Mixture Models [26.847612684502998]
We study the statistical performance of kernel-based clustering algorithms under non-parametric mixture models.
We establish a key equivalence between kernel-based data-clustering and kernel density-based clustering.
arXiv Detail & Related papers (2021-10-18T17:23:54Z) - Kernel k-Means, By All Means: Algorithms and Strong Consistency [21.013169939337583]
Kernel $k$ clustering is a powerful tool for unsupervised learning of non-linear data.
In this paper, we generalize results leveraging a general family of means to combat sub-optimal local solutions.
Our algorithm makes use of majorization-minimization (MM) to better solve this non-linear separation problem.
arXiv Detail & Related papers (2020-11-12T16:07:18Z) - The Impact of Isolation Kernel on Agglomerative Hierarchical Clustering
Algorithms [12.363083467305787]
Agglomerative hierarchical clustering (AHC) is one of the popular clustering approaches.
Existing AHC methods, which are based on a distance measure, have difficulty in identifying adjacent clusters with varied densities.
We show that the use of a data-dependent kernel (instead of distance or existing kernel) provides an effective means to address it.
arXiv Detail & Related papers (2020-10-12T06:18:38Z) - Kernel learning approaches for summarising and combining posterior
similarity matrices [68.8204255655161]
We build upon the notion of the posterior similarity matrix (PSM) in order to suggest new approaches for summarising the output of MCMC algorithms for Bayesian clustering models.
A key contribution of our work is the observation that PSMs are positive semi-definite, and hence can be used to define probabilistically-motivated kernel matrices.
arXiv Detail & Related papers (2020-09-27T14:16:14Z) - SimpleMKKM: Simple Multiple Kernel K-means [49.500663154085586]
We propose a simple yet effective multiple kernel clustering algorithm, termed simple multiple kernel k-means (SimpleMKKM)
Our criterion is given by an intractable minimization-maximization problem in the kernel coefficient and clustering partition matrix.
We theoretically analyze the performance of SimpleMKKM in terms of its clustering generalization error.
arXiv Detail & Related papers (2020-05-11T10:06:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.