Multiple Kernel Clustering with Dual Noise Minimization
- URL: http://arxiv.org/abs/2207.06041v1
- Date: Wed, 13 Jul 2022 08:37:42 GMT
- Title: Multiple Kernel Clustering with Dual Noise Minimization
- Authors: Junpu Zhang and Liang Li and Siwei Wang and Jiyuan Liu and Yue Liu and
Xinwang Liu and En Zhu
- Abstract summary: Multiple kernel clustering (MKC) aims to group data by integrating complementary information from base kernels.
In this paper, we rigorously define dual noise and propose a novel parameter-free MKC algorithm by minimizing them.
We observe that dual noise will pollute the block diagonal structures and incur the degeneration of clustering performance, and C-noise exhibits stronger destruction than N-noise.
- Score: 56.009011016367744
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Clustering is a representative unsupervised method widely applied in
multi-modal and multi-view scenarios. Multiple kernel clustering (MKC) aims to
group data by integrating complementary information from base kernels. As a
representative, late fusion MKC first decomposes the kernels into orthogonal
partition matrices, then learns a consensus one from them, achieving promising
performance recently. However, these methods fail to consider the noise inside
the partition matrix, preventing further improvement of clustering performance.
We discover that the noise can be disassembled into separable dual parts, i.e.
N-noise and C-noise (Null space noise and Column space noise). In this paper,
we rigorously define dual noise and propose a novel parameter-free MKC
algorithm by minimizing them. To solve the resultant optimization problem, we
design an efficient two-step iterative strategy. To our best knowledge, it is
the first time to investigate dual noise within the partition in the kernel
space. We observe that dual noise will pollute the block diagonal structures
and incur the degeneration of clustering performance, and C-noise exhibits
stronger destruction than N-noise. Owing to our efficient mechanism to minimize
dual noise, the proposed algorithm surpasses the recent methods by large
margins.
Related papers
- Multiple Kernel Clustering via Local Regression Integration [4.856913393644719]
Multiple kernel methods less consider the intrinsic manifold structure of multiple kernel data.
This paper first presents the clustering method via kernelized local regression (CKLR)
We then extend it to perform clustering via the multiple kernel local regression (CMKLR)
arXiv Detail & Related papers (2024-10-20T06:26:29Z) - MOKD: Cross-domain Finetuning for Few-shot Classification via Maximizing Optimized Kernel Dependence [97.93517982908007]
In cross-domain few-shot classification, NCC aims to learn representations to construct a metric space where few-shot classification can be performed.
In this paper, we find that there exist high similarities between NCC-learned representations of two samples from different classes.
We propose a bi-level optimization framework, emphmaximizing optimized kernel dependence (MOKD) to learn a set of class-specific representations that match the cluster structures indicated by labeled data.
arXiv Detail & Related papers (2024-05-29T05:59:52Z) - Towards Robust Few-shot Point Cloud Semantic Segmentation [57.075074484313]
Few-shot point cloud semantic segmentation aims to train a model to quickly adapt to new unseen classes with only a handful of support set samples.
We propose a Component-level Clean Noise Separation (CCNS) representation learning to learn discriminative feature representations.
We also propose a Multi-scale Degree-based Noise Suppression (MDNS) scheme to remove the noisy shots from the support set.
arXiv Detail & Related papers (2023-09-20T11:40:10Z) - POCKET: Pruning Random Convolution Kernels for Time Series Classification from a Feature Selection Perspective [8.359327841946852]
A time series classification model, POCKET, is designed to efficiently prune redundant kernels.
POCKET prunes up to 60% of kernels without a significant reduction in accuracy and performs 11$times$ faster than its counterparts.
Experimental results on diverse time series datasets show that POCKET prunes up to 60% of kernels without a significant reduction in accuracy and performs 11$times$ faster than its counterparts.
arXiv Detail & Related papers (2023-09-15T16:03:23Z) - Local Sample-weighted Multiple Kernel Clustering with Consensus
Discriminative Graph [73.68184322526338]
Multiple kernel clustering (MKC) is committed to achieving optimal information fusion from a set of base kernels.
This paper proposes a novel local sample-weighted multiple kernel clustering model.
Experimental results demonstrate that our LSWMKC possesses better local manifold representation and outperforms existing kernel or graph-based clustering algo-rithms.
arXiv Detail & Related papers (2022-07-05T05:00:38Z) - Learning "best" kernels from data in Gaussian process regression. With
application to aerodynamics [0.4588028371034406]
We introduce algorithms to select/design kernels in Gaussian process regression/kriging surrogate modeling techniques.
A first class of algorithms is kernel flow, which was introduced in a context of classification in machine learning.
A second class of algorithms is called spectral kernel ridge regression, and aims at selecting a "best" kernel such that the norm of the function to be approximated is minimal.
arXiv Detail & Related papers (2022-06-03T07:50:54Z) - Learning Halfspaces with Tsybakov Noise [50.659479930171585]
We study the learnability of halfspaces in the presence of Tsybakov noise.
We give an algorithm that achieves misclassification error $epsilon$ with respect to the true halfspace.
arXiv Detail & Related papers (2020-06-11T14:25:02Z) - SimpleMKKM: Simple Multiple Kernel K-means [49.500663154085586]
We propose a simple yet effective multiple kernel clustering algorithm, termed simple multiple kernel k-means (SimpleMKKM)
Our criterion is given by an intractable minimization-maximization problem in the kernel coefficient and clustering partition matrix.
We theoretically analyze the performance of SimpleMKKM in terms of its clustering generalization error.
arXiv Detail & Related papers (2020-05-11T10:06:40Z) - Provable Noisy Sparse Subspace Clustering using Greedy Neighbor
Selection: A Coherence-Based Perspective [18.888312436971187]
We derive coherence-based sufficient conditions guaranteeing correct neighbor identification using MP/OMP.
A striking finding is that, when the ground truth subspaces are well-separated from each other and noise is not large, MP-based iterations, while enjoying lower algorithmic complexity, yield smaller perturbation of residuals.
arXiv Detail & Related papers (2020-02-02T14:28:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.