Boost Decentralized Federated Learning in Vehicular Networks by
Diversifying Data Sources
- URL: http://arxiv.org/abs/2209.01750v1
- Date: Mon, 5 Sep 2022 04:01:41 GMT
- Title: Boost Decentralized Federated Learning in Vehicular Networks by
Diversifying Data Sources
- Authors: Dongyuan Su, Yipeng Zhou, Laizhong Cui
- Abstract summary: We propose the DFL-DDS (DFL with diversified Data Sources) algorithm to diversify data sources in DFL.
Specifically, each vehicle maintains a state vector to record the contribution weight of each data source to its model.
To boost the convergence of DFL, a vehicle tunes the aggregation weight of each data source by minimizing the KL divergence of its state vector.
- Score: 16.342217928468227
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recently, federated learning (FL) has received intensive research because of
its ability in preserving data privacy for scattered clients to collaboratively
train machine learning models. Commonly, a parameter server (PS) is deployed
for aggregating model parameters contributed by different clients.
Decentralized federated learning (DFL) is upgraded from FL which allows clients
to aggregate model parameters with their neighbours directly. DFL is
particularly feasible for vehicular networks as vehicles communicate with each
other in a vehicle-to-vehicle (V2V) manner. However, due to the restrictions of
vehicle routes and communication distances, it is hard for individual vehicles
to sufficiently exchange models with others. Data sources contributing to
models on individual vehicles may not diversified enough resulting in poor
model accuracy. To address this problem, we propose the DFL-DDS (DFL with
diversified Data Sources) algorithm to diversify data sources in DFL.
Specifically, each vehicle maintains a state vector to record the contribution
weight of each data source to its model. The Kullback-Leibler (KL) divergence
is adopted to measure the diversity of a state vector. To boost the convergence
of DFL, a vehicle tunes the aggregation weight of each data source by
minimizing the KL divergence of its state vector, and its effectiveness in
diversifying data sources can be theoretically proved. Finally, the superiority
of DFL-DDS is evaluated by extensive experiments (with MNIST and CIFAR-10
datasets) which demonstrate that DFL-DDS can accelerate the convergence of DFL
and improve the model accuracy significantly compared with state-of-the-art
baselines.
Related papers
- MultiConfederated Learning: Inclusive Non-IID Data handling with Decentralized Federated Learning [1.2726316791083532]
Federated Learning (FL) has emerged as a prominent privacy-preserving technique for enabling use cases like confidential clinical machine learning.
FL operates by aggregating models trained by remote devices which owns the data.
We propose MultiConfederated Learning: a decentralized FL framework which is designed to handle non-IID data.
arXiv Detail & Related papers (2024-04-20T16:38:26Z) - Adaptive Model Pruning and Personalization for Federated Learning over
Wireless Networks [72.59891661768177]
Federated learning (FL) enables distributed learning across edge devices while protecting data privacy.
We consider a FL framework with partial model pruning and personalization to overcome these challenges.
This framework splits the learning model into a global part with model pruning shared with all devices to learn data representations and a personalized part to be fine-tuned for a specific device.
arXiv Detail & Related papers (2023-09-04T21:10:45Z) - NeFL: Nested Model Scaling for Federated Learning with System Heterogeneous Clients [44.89061671579694]
Federated learning (FL) enables distributed training while preserving data privacy, but stragglers-slow or incapable clients-can significantly slow down the total training time and degrade performance.
We propose nested federated learning (NeFL), a framework that efficiently divides deep neural networks into submodels using both depthwise and widthwise scaling.
NeFL achieves performance gain, especially for the worst-case submodel compared to baseline approaches.
arXiv Detail & Related papers (2023-08-15T13:29:14Z) - Tram-FL: Routing-based Model Training for Decentralized Federated
Learning [2.8558942410497066]
We propose Tram-FL, a novel DFL method, which progressively refines a global model by transferring it sequentially amongst nodes.
We also introduce a dynamic model routing algorithm for optimal route selection, aimed at enhancing model precision with minimal forwarding.
Our experiments using MNIST, CIFAR-10, and IMDb datasets demonstrate that Tram-FL with the proposed routing delivers high model accuracy under non-IID conditions.
arXiv Detail & Related papers (2023-08-09T07:51:07Z) - Communication Resources Constrained Hierarchical Federated Learning for
End-to-End Autonomous Driving [67.78611905156808]
This paper proposes an optimization-based Communication Resource Constrained Hierarchical Federated Learning framework.
Results show that the proposed CRCHFL both accelerates the convergence rate and enhances the generalization of federated learning autonomous driving model.
arXiv Detail & Related papers (2023-06-28T12:44:59Z) - GPT-FL: Generative Pre-trained Model-Assisted Federated Learning [40.522864349440674]
GPT-FL is a generative pre-trained model-assisted federated learning framework.
It consistently outperforms state-of-the-art FL methods in terms of model test accuracy, communication efficiency, and client sampling efficiency.
arXiv Detail & Related papers (2023-06-03T22:57:59Z) - Asynchronous Federated Learning for Edge-assisted Vehicular Networks [7.624367655819205]
Vehicular networks enable vehicles to support real-time vehicular applications through training data.
For the traditional federated learning (FL), vehicles train the data locally to obtain a local model and then upload the local model to the RSU to update the global model.
The traditional FL updates the global model synchronously, i.e., the RSU needs to wait for all vehicles to upload their models for the global model updating.
It is necessary to propose an asynchronous federated learning (AFL) to solve this problem, where the RSU updates the global model once it receives a local model from a vehicle
arXiv Detail & Related papers (2022-08-03T08:05:02Z) - FedDM: Iterative Distribution Matching for Communication-Efficient
Federated Learning [87.08902493524556]
Federated learning(FL) has recently attracted increasing attention from academia and industry.
We propose FedDM to build the global training objective from multiple local surrogate functions.
In detail, we construct synthetic sets of data on each client to locally match the loss landscape from original data.
arXiv Detail & Related papers (2022-07-20T04:55:18Z) - Parallel Successive Learning for Dynamic Distributed Model Training over
Heterogeneous Wireless Networks [50.68446003616802]
Federated learning (FedL) has emerged as a popular technique for distributing model training over a set of wireless devices.
We develop parallel successive learning (PSL), which expands the FedL architecture along three dimensions.
Our analysis sheds light on the notion of cold vs. warmed up models, and model inertia in distributed machine learning.
arXiv Detail & Related papers (2022-02-07T05:11:01Z) - Over-the-Air Federated Learning from Heterogeneous Data [107.05618009955094]
Federated learning (FL) is a framework for distributed learning of centralized models.
We develop a Convergent OTA FL (COTAF) algorithm which enhances the common local gradient descent (SGD) FL algorithm.
We numerically show that the precoding induced by COTAF notably improves the convergence rate and the accuracy of models trained via OTA FL.
arXiv Detail & Related papers (2020-09-27T08:28:25Z) - Ensemble Distillation for Robust Model Fusion in Federated Learning [72.61259487233214]
Federated Learning (FL) is a machine learning setting where many devices collaboratively train a machine learning model.
In most of the current training schemes the central model is refined by averaging the parameters of the server model and the updated parameters from the client side.
We propose ensemble distillation for model fusion, i.e. training the central classifier through unlabeled data on the outputs of the models from the clients.
arXiv Detail & Related papers (2020-06-12T14:49:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.