Explaining Kernel Clustering via Decision Trees
- URL: http://arxiv.org/abs/2402.09881v1
- Date: Thu, 15 Feb 2024 11:08:23 GMT
- Title: Explaining Kernel Clustering via Decision Trees
- Authors: Maximilian Fleissner, Leena Chennuru Vankadara, Debarghya
Ghoshdastidar
- Abstract summary: We investigate interpretable kernel clustering, and propose algorithms that construct decision trees to approximate partitions induced by kernel k-means.
We build on previous work on explainable k-means and demonstrate how a suitable choice of features allows preserving interpretability without sacrificing approximation guarantees on the interpretable model.
- Score: 10.504801686625129
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Despite the growing popularity of explainable and interpretable machine
learning, there is still surprisingly limited work on inherently interpretable
clustering methods. Recently, there has been a surge of interest in explaining
the classic k-means algorithm, leading to efficient algorithms that approximate
k-means clusters using axis-aligned decision trees. However, interpretable
variants of k-means have limited applicability in practice, where more flexible
clustering methods are often needed to obtain useful partitions of the data. In
this work, we investigate interpretable kernel clustering, and propose
algorithms that construct decision trees to approximate the partitions induced
by kernel k-means, a nonlinear extension of k-means. We further build on
previous work on explainable k-means and demonstrate how a suitable choice of
features allows preserving interpretability without sacrificing approximation
guarantees on the interpretable model.
Related papers
- Self-Supervised Graph Embedding Clustering [70.36328717683297]
K-means one-step dimensionality reduction clustering method has made some progress in addressing the curse of dimensionality in clustering tasks.
We propose a unified framework that integrates manifold learning with K-means, resulting in the self-supervised graph embedding framework.
arXiv Detail & Related papers (2024-09-24T08:59:51Z) - Wide Gaps and Clustering Axioms [0.0]
k-means is in conflict with Kleinberg's axiomatic system for distance based clustering algorithms.
We introduce two new clusterability properties, variational k-separability and residual k-separability.
We reconcile k-means with Kleinberg's axiomatic framework in Euclidean and non-Euclidean settings.
arXiv Detail & Related papers (2023-08-07T10:43:48Z) - Rethinking k-means from manifold learning perspective [122.38667613245151]
We present a new clustering algorithm which directly detects clusters of data without mean estimation.
Specifically, we construct distance matrix between data points by Butterworth filter.
To well exploit the complementary information embedded in different views, we leverage the tensor Schatten p-norm regularization.
arXiv Detail & Related papers (2023-05-12T03:01:41Z) - Sketch-and-solve approaches to k-means clustering by semidefinite
programming [14.930208990741132]
We introduce a sketch-and-solve approach to speed up the Peng-Wei semidefinite relaxation of k-means clustering.
If the data is appropriately separated we identify the k-means optimal clustering.
Otherwise, our approach provides a high-confidence lower bound on the optimal k-means value.
arXiv Detail & Related papers (2022-11-28T19:51:30Z) - A One-shot Framework for Distributed Clustered Learning in Heterogeneous
Environments [54.172993875654015]
The paper proposes a family of communication efficient methods for distributed learning in heterogeneous environments.
One-shot approach, based on local computations at the users and a clustering based aggregation step at the server is shown to provide strong learning guarantees.
For strongly convex problems it is shown that, as long as the number of data points per user is above a threshold, the proposed approach achieves order-optimal mean-squared error rates in terms of the sample size.
arXiv Detail & Related papers (2022-09-22T09:04:10Z) - Perfect Spectral Clustering with Discrete Covariates [68.8204255655161]
We propose a spectral algorithm that achieves perfect clustering with high probability on a class of large, sparse networks.
Our method is the first to offer a guarantee of consistent latent structure recovery using spectral clustering.
arXiv Detail & Related papers (2022-05-17T01:41:06Z) - Adaptive Explicit Kernel Minkowski Weighted K-means [1.3535770763481905]
The kernel K-means, which extends K-means into the kernel space, is able to capture nonlinear structures and identify arbitrarily shaped clusters.
This paper proposes a method to combine the advantages of the linear and nonlinear approaches by using driven corresponding approximate finite-dimensional feature maps.
arXiv Detail & Related papers (2020-12-04T19:14:09Z) - Scalable Hierarchical Agglomerative Clustering [65.66407726145619]
Existing scalable hierarchical clustering methods sacrifice quality for speed.
We present a scalable, agglomerative method for hierarchical clustering that does not sacrifice quality and scales to billions of data points.
arXiv Detail & Related papers (2020-10-22T15:58:35Z) - Differentially Private Clustering: Tight Approximation Ratios [57.89473217052714]
We give efficient differentially private algorithms for basic clustering problems.
Our results imply an improved algorithm for the Sample and Aggregate privacy framework.
One of the tools used in our 1-Cluster algorithm can be employed to get a faster quantum algorithm for ClosestPair in a moderate number of dimensions.
arXiv Detail & Related papers (2020-08-18T16:22:06Z) - A generalized Bayes framework for probabilistic clustering [3.3194866396158]
Loss-based clustering methods, such as k-means and its variants, are standard tools for finding groups in data.
Model-based clustering based on mixture models provides an alternative, but such methods face computational problems and large sensitivity to the choice of kernel.
This article proposes a generalized Bayes framework that bridges between these two paradigms through the use of Gibbs posteriors.
arXiv Detail & Related papers (2020-06-09T18:49:32Z) - Simple and Scalable Sparse k-means Clustering via Feature Ranking [14.839931533868176]
We propose a novel framework for sparse k-means clustering that is intuitive, simple to implement, and competitive with state-of-the-art algorithms.
Our core method readily generalizes to several task-specific algorithms such as clustering on subsets of attributes and in partially observed data settings.
arXiv Detail & Related papers (2020-02-20T02:41:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.