Federated K-Means Clustering via Dual Decomposition-based Distributed
Optimization
- URL: http://arxiv.org/abs/2307.13267v1
- Date: Tue, 25 Jul 2023 05:34:50 GMT
- Title: Federated K-Means Clustering via Dual Decomposition-based Distributed
Optimization
- Authors: Vassilios Yfantis, Achim Wagner, Martin Ruskowski
- Abstract summary: This paper shows how dual decomposition can be applied for distributed training of $ K $-means clustering problems.
The training can be performed in a distributed manner by splitting the data across different nodes and linking these nodes through consensus constraints.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The use of distributed optimization in machine learning can be motivated
either by the resulting preservation of privacy or the increase in
computational efficiency. On the one hand, training data might be stored across
multiple devices. Training a global model within a network where each node only
has access to its confidential data requires the use of distributed algorithms.
Even if the data is not confidential, sharing it might be prohibitive due to
bandwidth limitations. On the other hand, the ever-increasing amount of
available data leads to large-scale machine learning problems. By splitting the
training process across multiple nodes its efficiency can be significantly
increased. This paper aims to demonstrate how dual decomposition can be applied
for distributed training of $ K $-means clustering problems. After an overview
of distributed and federated machine learning, the mixed-integer quadratically
constrained programming-based formulation of the $ K $-means clustering
training problem is presented. The training can be performed in a distributed
manner by splitting the data across different nodes and linking these nodes
through consensus constraints. Finally, the performance of the subgradient
method, the bundle trust method, and the quasi-Newton dual ascent algorithm are
evaluated on a set of benchmark problems. While the mixed-integer
programming-based formulation of the clustering problems suffers from weak
integer relaxations, the presented approach can potentially be used to enable
an efficient solution in the future, both in a central and distributed setting.
Related papers
- Probabilistic Contrastive Learning for Long-Tailed Visual Recognition [78.70453964041718]
Longtailed distributions frequently emerge in real-world data, where a large number of minority categories contain a limited number of samples.
Recent investigations have revealed that supervised contrastive learning exhibits promising potential in alleviating the data imbalance.
We propose a novel probabilistic contrastive (ProCo) learning algorithm that estimates the data distribution of the samples from each class in the feature space.
arXiv Detail & Related papers (2024-03-11T13:44:49Z) - Large Scale Constrained Clustering With Reinforcement Learning [1.3597551064547502]
Given a network, allocating resources at clusters level, rather than at each node, enhances efficiency in resource allocation and usage.
We propose an approach to solve this constrained clustering problem via reinforcement learning.
In the results section, we show that our algorithm finds near optimal solutions, even for large scale instances.
arXiv Detail & Related papers (2024-02-15T18:27:18Z) - Ravnest: Decentralized Asynchronous Training on Heterogeneous Devices [0.0]
Ravnest facilitates decentralized training by efficiently organizing compute nodes into clusters.
We have framed our asynchronous SGD loss function as a block structured optimization problem with delayed updates.
arXiv Detail & Related papers (2024-01-03T13:07:07Z) - Communication-Efficient Decentralized Federated Learning via One-Bit
Compressive Sensing [52.402550431781805]
Decentralized federated learning (DFL) has gained popularity due to its practicality across various applications.
Compared to the centralized version, training a shared model among a large number of nodes in DFL is more challenging.
We develop a novel algorithm based on the framework of the inexact alternating direction method (iADM)
arXiv Detail & Related papers (2023-08-31T12:22:40Z) - Hard Regularization to Prevent Deep Online Clustering Collapse without
Data Augmentation [65.268245109828]
Online deep clustering refers to the joint use of a feature extraction network and a clustering model to assign cluster labels to each new data point or batch as it is processed.
While faster and more versatile than offline methods, online clustering can easily reach the collapsed solution where the encoder maps all inputs to the same point and all are put into a single cluster.
We propose a method that does not require data augmentation, and that, differently from existing methods, regularizes the hard assignments.
arXiv Detail & Related papers (2023-03-29T08:23:26Z) - Weight Divergence Driven Divide-and-Conquer Approach for Optimal
Federated Learning from non-IID Data [0.0]
Federated Learning allows training of data stored in distributed devices without the need for centralizing training data.
We propose a novel Divide-and-Conquer training methodology that enables the use of the popular FedAvg aggregation algorithm.
arXiv Detail & Related papers (2021-06-28T09:34:20Z) - Coded Stochastic ADMM for Decentralized Consensus Optimization with Edge
Computing [113.52575069030192]
Big data, including applications with high security requirements, are often collected and stored on multiple heterogeneous devices, such as mobile devices, drones and vehicles.
Due to the limitations of communication costs and security requirements, it is of paramount importance to extract information in a decentralized manner instead of aggregating data to a fusion center.
We consider the problem of learning model parameters in a multi-agent system with data locally processed via distributed edge nodes.
A class of mini-batch alternating direction method of multipliers (ADMM) algorithms is explored to develop the distributed learning model.
arXiv Detail & Related papers (2020-10-02T10:41:59Z) - A Low Complexity Decentralized Neural Net with Centralized Equivalence
using Layer-wise Learning [49.15799302636519]
We design a low complexity decentralized learning algorithm to train a recently proposed large neural network in distributed processing nodes (workers)
In our setup, the training data is distributed among the workers but is not shared in the training process due to privacy and security concerns.
We show that it is possible to achieve equivalent learning performance as if the data is available in a single place.
arXiv Detail & Related papers (2020-09-29T13:08:12Z) - Learning while Respecting Privacy and Robustness to Distributional
Uncertainties and Adversarial Data [66.78671826743884]
The distributionally robust optimization framework is considered for training a parametric model.
The objective is to endow the trained model with robustness against adversarially manipulated input data.
Proposed algorithms offer robustness with little overhead.
arXiv Detail & Related papers (2020-07-07T18:25:25Z) - An Efficient Framework for Clustered Federated Learning [26.24231986590374]
We address the problem of federated learning (FL) where users are distributed into clusters.
We propose the Iterative Federated Clustering Algorithm (IFCA)
We show that our algorithm is efficient in non- partitioned problems such as neural networks.
arXiv Detail & Related papers (2020-06-07T08:48:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.