Incremental Affinity Propagation based on Cluster Consolidation and
Stratification
- URL: http://arxiv.org/abs/2401.14439v1
- Date: Thu, 25 Jan 2024 14:20:00 GMT
- Title: Incremental Affinity Propagation based on Cluster Consolidation and
Stratification
- Authors: Silvana Castano, Alfio Ferrara, Stefano Montanelli, Francesco Periti
- Abstract summary: We propose A-Posteriori affinity Propagation (APP) to achieve faithfulness and forgetfulness.
APP enforces incremental clustering where i) new arriving objects are dynamically consolidated into previous clusters without the need to re-execute clustering over the entire dataset of objects.
Experimental results show that APP achieves comparable clustering performance while enforcing scalability at the same time.
- Score: 2.048226951354646
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Modern data mining applications require to perform incremental clustering
over dynamic datasets by tracing temporal changes over the resulting clusters.
In this paper, we propose A-Posteriori affinity Propagation (APP), an
incremental extension of Affinity Propagation (AP) based on cluster
consolidation and cluster stratification to achieve faithfulness and
forgetfulness. APP enforces incremental clustering where i) new arriving
objects are dynamically consolidated into previous clusters without the need to
re-execute clustering over the entire dataset of objects, and ii) a faithful
sequence of clustering results is produced and maintained over time, while
allowing to forget obsolete clusters with decremental learning functionalities.
Four popular labeled datasets are used to test the performance of APP with
respect to benchmark clustering performances obtained by conventional AP and
Incremental Affinity Propagation based on Nearest neighbor Assignment (IAPNA)
algorithms. Experimental results show that APP achieves comparable clustering
performance while enforcing scalability at the same time.
Related papers
- Adaptive Self-supervised Robust Clustering for Unstructured Data with Unknown Cluster Number [12.926206811876174]
We introduce a novel self-supervised deep clustering approach tailored for unstructured data, termed Adaptive Self-supervised Robust Clustering (ASRC)
ASRC adaptively learns the graph structure and edge weights to capture both local and global structural information.
ASRC even outperforms methods that rely on prior knowledge of the number of clusters, highlighting its effectiveness in addressing the challenges of clustering unstructured data.
arXiv Detail & Related papers (2024-07-29T15:51:09Z) - A3S: A General Active Clustering Method with Pairwise Constraints [66.74627463101837]
A3S features strategic active clustering adjustment on the initial cluster result, which is obtained by an adaptive clustering algorithm.
In extensive experiments across diverse real-world datasets, A3S achieves desired results with significantly fewer human queries.
arXiv Detail & Related papers (2024-07-14T13:37:03Z) - Towards Efficient and Effective Deep Clustering with Dynamic Grouping
and Prototype Aggregation [4.550555443103878]
We present a novel end-to-end deep clustering framework with dynamic grouping and prototype aggregation, termed as DigPro.
Specifically, the proposed dynamic grouping extends contrastive learning from instance-level to group-level, which is effective and efficient for timely updating groups.
With an expectation-maximization framework, DigPro simultaneously takes advantage of compact intra-cluster connections, well-separated clusters, and efficient group updating during the self-supervised training.
arXiv Detail & Related papers (2024-01-24T16:45:42Z) - Reinforcement Graph Clustering with Unknown Cluster Number [91.4861135742095]
We propose a new deep graph clustering method termed Reinforcement Graph Clustering.
In our proposed method, cluster number determination and unsupervised representation learning are unified into a uniform framework.
In order to conduct feedback actions, the clustering-oriented reward function is proposed to enhance the cohesion of the same clusters and separate the different clusters.
arXiv Detail & Related papers (2023-08-13T18:12:28Z) - Instance-Optimal Cluster Recovery in the Labeled Stochastic Block Model [79.46465138631592]
We devise an efficient algorithm that recovers clusters using the observed labels.
We present Instance-Adaptive Clustering (IAC), the first algorithm whose performance matches these lower bounds both in expectation and with high probability.
arXiv Detail & Related papers (2023-06-18T08:46:06Z) - Mitigating shortage of labeled data using clustering-based active
learning with diversity exploration [3.312798619476657]
We propose a clustering-based active learning framework, namely Active Learning using a Clustering-based Sampling.
A bi-cluster boundary-based sample query procedure is introduced to improve the learning performance for classifying highly overlapped classes.
arXiv Detail & Related papers (2022-07-06T20:53:28Z) - Self-supervised Contrastive Attributed Graph Clustering [110.52694943592974]
We propose a novel attributed graph clustering network, namely Self-supervised Contrastive Attributed Graph Clustering (SCAGC)
In SCAGC, by leveraging inaccurate clustering labels, a self-supervised contrastive loss, are designed for node representation learning.
For the OOS nodes, SCAGC can directly calculate their clustering labels.
arXiv Detail & Related papers (2021-10-15T03:25:28Z) - You Never Cluster Alone [150.94921340034688]
We extend the mainstream contrastive learning paradigm to a cluster-level scheme, where all the data subjected to the same cluster contribute to a unified representation.
We define a set of categorical variables as clustering assignment confidence, which links the instance-level learning track with the cluster-level one.
By reparametrizing the assignment variables, TCC is trained end-to-end, requiring no alternating steps.
arXiv Detail & Related papers (2021-06-03T14:59:59Z) - Scalable Hierarchical Agglomerative Clustering [65.66407726145619]
Existing scalable hierarchical clustering methods sacrifice quality for speed.
We present a scalable, agglomerative method for hierarchical clustering that does not sacrifice quality and scales to billions of data points.
arXiv Detail & Related papers (2020-10-22T15:58:35Z) - Evolutionary Clustering via Message Passing [21.366647840787365]
We introduce evolutionary affinity propagation (EAP), an evolutionary clustering algorithm that groups data points by exchanging messages on a factor graph.
EAP promotes temporal smoothness of the solution to clustering time-evolving data by linking the nodes of the factor graph that are associated with adjacent data snapshots.
EAP determines the number of clusters and tracks them automatically.
arXiv Detail & Related papers (2019-12-27T03:09:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.