Enhanced Decentralized Federated Learning based on Consensus in
Connected Vehicles
- URL: http://arxiv.org/abs/2209.10722v1
- Date: Thu, 22 Sep 2022 01:21:23 GMT
- Title: Enhanced Decentralized Federated Learning based on Consensus in
Connected Vehicles
- Authors: Xiaoyan Liu, Zehui Dong, Zhiwei Xu, Siyuan Liu, Jie Tian
- Abstract summary: Federated learning (FL) is emerging as a new paradigm to train machine learning (ML) models in distributed systems.
We introduce C-DFL (Consensus based Decentralized Federated Learning) to tackle federated learning on connected vehicles.
- Score: 14.80476265018825
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Advanced researches on connected vehicles have recently targeted to the
integration of vehicle-to-everything (V2X) networks with Machine Learning (ML)
tools and distributed decision making. Federated learning (FL) is emerging as a
new paradigm to train machine learning (ML) models in distributed systems,
including vehicles in V2X networks. Rather than sharing and uploading the
training data to the server, the updating of model parameters (e.g., neural
networks' weights and biases) is applied by large populations of interconnected
vehicles, acting as local learners. Despite these benefits, the limitation of
existing approaches is the centralized optimization which relies on a server
for aggregation and fusion of local parameters, leading to the drawback of a
single point of failure and scaling issues for increasing V2X network size.
Meanwhile, in intelligent transport scenarios, data collected from onboard
sensors are redundant, which degrades the performance of aggregation. To tackle
these problems, we explore a novel idea of decentralized data processing and
introduce a federated learning framework for in-network vehicles,
C-DFL(Consensus based Decentralized Federated Learning), to tackle federated
learning on connected vehicles and improve learning quality. Extensive
simulations have been implemented to evaluate the performance of C-DFL, that
demonstrates C-DFL outperforms the performance of conventional methods in all
cases.
Related papers
- Adaptive and Parallel Split Federated Learning in Vehicular Edge Computing [6.004901615052089]
Vehicular edge intelligence (VEI) is a promising paradigm for enabling future intelligent transportation systems.
Federated learning (FL) is one of the fundamental technologies facilitating collaborative model training locally and aggregation.
We develop an Adaptive Split Federated Learning scheme for Vehicular Edge Computing (ASFV)
arXiv Detail & Related papers (2024-05-29T02:34:38Z) - Scheduling and Communication Schemes for Decentralized Federated
Learning [0.31410859223862103]
A decentralized federated learning (DFL) model with the gradient descent (SGD) algorithm has been introduced.
Three scheduling policies for DFL have been proposed for communications between the clients and the parallel servers.
Results show that the proposed scheduling polices have an impact both on the speed of convergence and in the final global model.
arXiv Detail & Related papers (2023-11-27T17:35:28Z) - Semi-Federated Learning: Convergence Analysis and Optimization of A
Hybrid Learning Framework [70.83511997272457]
We propose a semi-federated learning (SemiFL) paradigm to leverage both the base station (BS) and devices for a hybrid implementation of centralized learning (CL) and FL.
We propose a two-stage algorithm to solve this intractable problem, in which we provide the closed-form solutions to the beamformers.
arXiv Detail & Related papers (2023-10-04T03:32:39Z) - Towards Cooperative Federated Learning over Heterogeneous Edge/Fog
Networks [49.19502459827366]
Federated learning (FL) has been promoted as a popular technique for training machine learning (ML) models over edge/fog networks.
Traditional implementations of FL have largely neglected the potential for inter-network cooperation.
We advocate for cooperative federated learning (CFL), a cooperative edge/fog ML paradigm built on device-to-device (D2D) and device-to-server (D2S) interactions.
arXiv Detail & Related papers (2023-03-15T04:41:36Z) - Time-sensitive Learning for Heterogeneous Federated Edge Intelligence [52.83633954857744]
We investigate real-time machine learning in a federated edge intelligence (FEI) system.
FEI systems exhibit heterogenous communication and computational resource distribution.
We propose a time-sensitive federated learning (TS-FL) framework to minimize the overall run-time for collaboratively training a shared ML model.
arXiv Detail & Related papers (2023-01-26T08:13:22Z) - Federated Deep Learning Meets Autonomous Vehicle Perception: Design and
Verification [168.67190934250868]
Federated learning empowered connected autonomous vehicle (FLCAV) has been proposed.
FLCAV preserves privacy while reducing communication and annotation costs.
It is challenging to determine the network resources and road sensor poses for multi-stage training.
arXiv Detail & Related papers (2022-06-03T23:55:45Z) - Multi-Edge Server-Assisted Dynamic Federated Learning with an Optimized
Floating Aggregation Point [51.47520726446029]
cooperative edge learning (CE-FL) is a distributed machine learning architecture.
We model the processes taken during CE-FL, and conduct analytical training.
We show the effectiveness of our framework with the data collected from a real-world testbed.
arXiv Detail & Related papers (2022-03-26T00:41:57Z) - Parallel Successive Learning for Dynamic Distributed Model Training over
Heterogeneous Wireless Networks [50.68446003616802]
Federated learning (FedL) has emerged as a popular technique for distributing model training over a set of wireless devices.
We develop parallel successive learning (PSL), which expands the FedL architecture along three dimensions.
Our analysis sheds light on the notion of cold vs. warmed up models, and model inertia in distributed machine learning.
arXiv Detail & Related papers (2022-02-07T05:11:01Z) - Federated Learning with Cooperating Devices: A Consensus Approach for
Massive IoT Networks [8.456633924613456]
Federated learning (FL) is emerging as a new paradigm to train machine learning models in distributed systems.
The paper proposes a fully distributed (or server-less) learning approach: the proposed FL algorithms leverage the cooperation of devices that perform data operations inside the network.
The approach lays the groundwork for integration of FL within 5G and beyond networks characterized by decentralized connectivity and computing.
arXiv Detail & Related papers (2019-12-27T15:16:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.