How Much Does It Cost to Train a Machine Learning Model over Distributed
Data Sources?
- URL: http://arxiv.org/abs/2209.07124v1
- Date: Thu, 15 Sep 2022 08:13:40 GMT
- Title: How Much Does It Cost to Train a Machine Learning Model over Distributed
Data Sources?
- Authors: Elia Guerra, Francesc Wilhelmi, Marco Miozzo, Paolo Dini
- Abstract summary: Federated learning allows devices to train a machine learning model without sharing their raw data.
Server-less FL approaches like gossip federated learning (GFL) and blockchain-enabled federated learning (BFL) have been proposed to mitigate these issues.
GFL is able to save the 18% of training time, the 68% of energy and the 51% of data to be shared with respect to the CFL solution, but it is not able to reach the level of accuracy of CFL.
BFL represents a viable solution for implementing decentralized learning with a higher level of security, at the cost of an extra energy usage and data sharing
- Score: 4.222078489059043
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Federated learning (FL) is one of the most appealing alternatives to the
standard centralized learning paradigm, allowing heterogeneous set of devices
to train a machine learning model without sharing their raw data. However, FL
requires a central server to coordinate the learning process, thus introducing
potential scalability and security issues. In the literature, server-less FL
approaches like gossip federated learning (GFL) and blockchain-enabled
federated learning (BFL) have been proposed to mitigate these issues. In this
work, we propose a complete overview of these three techniques proposing a
comparison according to an integral set of performance indicators, including
model accuracy, time complexity, communication overhead, convergence time and
energy consumption. An extensive simulation campaign permits to draw a
quantitative analysis. In particular, GFL is able to save the 18% of training
time, the 68% of energy and the 51% of data to be shared with respect to the
CFL solution, but it is not able to reach the level of accuracy of CFL. On the
other hand, BFL represents a viable solution for implementing decentralized
learning with a higher level of security, at the cost of an extra energy usage
and data sharing. Finally, we identify open issues on the two decentralized
federated learning implementations and provide insights on potential extensions
and possible research directions on this new research field.
Related papers
- Decentralized Personalized Federated Learning based on a Conditional Sparse-to-Sparser Scheme [5.5058010121503]
Decentralized Federated Learning (DFL) has become popular due to its robustness and avoidance of centralized coordination.
We propose a novel textitsparse-to-sparser training scheme: DA-DPFL.
Our experiments showcase that DA-DPFL substantially outperforms DFL baselines in test accuracy, while achieving up to $5$ times reduction in energy costs.
arXiv Detail & Related papers (2024-04-24T16:03:34Z) - Enhancing Trust and Privacy in Distributed Networks: A Comprehensive Survey on Blockchain-based Federated Learning [51.13534069758711]
Decentralized approaches like blockchain offer a compelling solution by implementing a consensus mechanism among multiple entities.
Federated Learning (FL) enables participants to collaboratively train models while safeguarding data privacy.
This paper investigates the synergy between blockchain's security features and FL's privacy-preserving model training capabilities.
arXiv Detail & Related papers (2024-03-28T07:08:26Z) - Communication Efficient ConFederated Learning: An Event-Triggered SAGA
Approach [67.27031215756121]
Federated learning (FL) is a machine learning paradigm that targets model training without gathering the local data over various data sources.
Standard FL, which employs a single server, can only support a limited number of users, leading to degraded learning capability.
In this work, we consider a multi-server FL framework, referred to as emphConfederated Learning (CFL) in order to accommodate a larger number of users.
arXiv Detail & Related papers (2024-02-28T03:27:10Z) - Federated Fine-Tuning of LLMs on the Very Edge: The Good, the Bad, the Ugly [62.473245910234304]
This paper takes a hardware-centric approach to explore how Large Language Models can be brought to modern edge computing systems.
We provide a micro-level hardware benchmark, compare the model FLOP utilization to a state-of-the-art data center GPU, and study the network utilization in realistic conditions.
arXiv Detail & Related papers (2023-10-04T20:27:20Z) - DeFL: Decentralized Weight Aggregation for Cross-silo Federated Learning [2.43923223501858]
Federated learning (FL) is an emerging promising paradigm of privacy-preserving machine learning (ML)
We propose DeFL, a novel decentralized weight aggregation framework for cross-silo FL.
DeFL eliminates the central server by aggregating weights on each participating node and weights of only the current training round are maintained and synchronized among all nodes.
arXiv Detail & Related papers (2022-08-01T13:36:49Z) - A Framework for Energy and Carbon Footprint Analysis of Distributed and
Federated Edge Learning [48.63610479916003]
This article breaks down and analyzes the main factors that influence the environmental footprint of distributed learning policies.
It models both vanilla and decentralized FL policies driven by consensus.
Results show that FL allows remarkable end-to-end energy savings (30%-40%) for wireless systems characterized by low bit/Joule efficiency.
arXiv Detail & Related papers (2021-03-18T16:04:42Z) - Wireless Communications for Collaborative Federated Learning [160.82696473996566]
Internet of Things (IoT) devices may not be able to transmit their collected data to a central controller for training machine learning models.
Google's seminal FL algorithm requires all devices to be directly connected with a central controller.
This paper introduces a novel FL framework, called collaborative FL (CFL), which enables edge devices to implement FL with less reliance on a central controller.
arXiv Detail & Related papers (2020-06-03T20:00:02Z) - Resource Management for Blockchain-enabled Federated Learning: A Deep
Reinforcement Learning Approach [54.29213445674221]
Federated Learning (BFL) enables mobile devices to collaboratively train neural network models required by a Machine Learning Model Owner (MLMO)
The issue of BFL is that the mobile devices have energy and CPU constraints that may reduce the system lifetime and training efficiency.
We propose to use the Deep Reinforcement Learning (DRL) to derive the optimal decisions for theO.
arXiv Detail & Related papers (2020-04-08T16:29:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.