Unsupervised collaborative learning using privileged information
- URL: http://arxiv.org/abs/2103.13145v1
- Date: Wed, 24 Mar 2021 12:43:49 GMT
- Title: Unsupervised collaborative learning using privileged information
- Authors: Yohan Foucade and Youn\`es Bennani
- Abstract summary: This article is dedicated to collaborative clustering based on the Learning Using Privileged Information paradigm.
A comparison between our algorithm and state of the art implementations shows improvement of the collaboration process using the proposed approach.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In the collaborative clustering framework, the hope is that by combining
several clustering solutions, each one with its own bias and imperfections, one
will get a better overall solution. The goal is that each local computation,
quite possibly applied to distinct data sets, benefits from the work done by
the other collaborators. This article is dedicated to collaborative clustering
based on the Learning Using Privileged Information paradigm. Local algorithms
weight incoming information at the level of each observation, depending on the
confidence level of the classification of that observation. A comparison
between our algorithm and state of the art implementations shows improvement of
the collaboration process using the proposed approach.
Related papers
- Data Similarity-Based One-Shot Clustering for Multi-Task Hierarchical Federated Learning [8.37314799155978]
We propose a one-shot clustering algorithm that can effectively identify and group users based on their data similarity.
Our proposed algorithm not only enhances the clustering process, but also overcomes challenges related to privacy concerns, communication overhead, and the need for prior knowledge about learning models or loss function behaviors.
arXiv Detail & Related papers (2024-10-03T17:51:21Z) - Discriminative Anchor Learning for Efficient Multi-view Clustering [59.11406089896875]
We propose discriminative anchor learning for multi-view clustering (DALMC)
We learn discriminative view-specific feature representations according to the original dataset.
We build anchors from different views based on these representations, which increase the quality of the shared anchor graph.
arXiv Detail & Related papers (2024-09-25T13:11:17Z) - Dynamically Weighted Federated k-Means [0.0]
Federated clustering enables multiple data sources to collaboratively cluster their data, maintaining decentralization and preserving privacy.
We introduce a novel federated clustering algorithm named Dynamically Weighted Federated k-means (DWF k-means) based on Lloyd's method for k-means clustering.
We conduct experiments on multiple datasets and data distribution settings to evaluate the performance of our algorithm in terms of clustering score, accuracy, and v-measure.
arXiv Detail & Related papers (2023-10-23T12:28:21Z) - Reinforcement Graph Clustering with Unknown Cluster Number [91.4861135742095]
We propose a new deep graph clustering method termed Reinforcement Graph Clustering.
In our proposed method, cluster number determination and unsupervised representation learning are unified into a uniform framework.
In order to conduct feedback actions, the clustering-oriented reward function is proposed to enhance the cohesion of the same clusters and separate the different clusters.
arXiv Detail & Related papers (2023-08-13T18:12:28Z) - Personalized Federated Learning with Feature Alignment and Classifier
Collaboration [13.320381377599245]
Data heterogeneity is one of the most challenging issues in federated learning.
One such approach in deep neural networks based tasks is employing a shared feature representation and learning a customized classifier head for each client.
In this work, we conduct explicit local-global feature alignment by leveraging global semantic knowledge for learning a better representation.
arXiv Detail & Related papers (2023-06-20T19:58:58Z) - MCoCo: Multi-level Consistency Collaborative Multi-view Clustering [15.743056561394612]
Multi-view clustering can explore consistent information from different views to guide clustering.
We propose a novel Multi-level Consistency Collaborative learning framework (MCoCo) for multi-view clustering.
arXiv Detail & Related papers (2023-02-26T16:08:53Z) - Decentralized adaptive clustering of deep nets is beneficial for client
collaboration [0.7012240324005975]
We study the problem of training personalized deep learning models in a decentralized peer-to-peer setting.
Our contribution is an algorithm which for each client finds beneficial collaborations based on a similarity estimate for the local task.
arXiv Detail & Related papers (2022-06-17T15:38:31Z) - On the Convergence of Clustered Federated Learning [57.934295064030636]
In a federated learning system, the clients, e.g. mobile devices and organization participants, usually have different personal preferences or behavior patterns.
This paper proposes a novel weighted client-based clustered FL algorithm to leverage the client's group and each client in a unified optimization framework.
arXiv Detail & Related papers (2022-02-13T02:39:19Z) - Leveraging Ensembles and Self-Supervised Learning for Fully-Unsupervised
Person Re-Identification and Text Authorship Attribution [77.85461690214551]
Learning from fully-unlabeled data is challenging in Multimedia Forensics problems, such as Person Re-Identification and Text Authorship Attribution.
Recent self-supervised learning methods have shown to be effective when dealing with fully-unlabeled data in cases where the underlying classes have significant semantic differences.
We propose a strategy to tackle Person Re-Identification and Text Authorship Attribution by enabling learning from unlabeled data even when samples from different classes are not prominently diverse.
arXiv Detail & Related papers (2022-02-07T13:08:11Z) - You Never Cluster Alone [150.94921340034688]
We extend the mainstream contrastive learning paradigm to a cluster-level scheme, where all the data subjected to the same cluster contribute to a unified representation.
We define a set of categorical variables as clustering assignment confidence, which links the instance-level learning track with the cluster-level one.
By reparametrizing the assignment variables, TCC is trained end-to-end, requiring no alternating steps.
arXiv Detail & Related papers (2021-06-03T14:59:59Z) - Graph Contrastive Clustering [131.67881457114316]
We propose a novel graph contrastive learning framework, which is then applied to the clustering task and we come up with the Graph Constrastive Clustering(GCC) method.
Specifically, on the one hand, the graph Laplacian based contrastive loss is proposed to learn more discriminative and clustering-friendly features.
On the other hand, a novel graph-based contrastive learning strategy is proposed to learn more compact clustering assignments.
arXiv Detail & Related papers (2021-04-03T15:32:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.