Data Heterogeneity-Robust Federated Learning via Group Client Selection
in Industrial IoT
- URL: http://arxiv.org/abs/2202.01512v1
- Date: Thu, 3 Feb 2022 10:48:17 GMT
- Title: Data Heterogeneity-Robust Federated Learning via Group Client Selection
in Industrial IoT
- Authors: Zonghang Li, Yihong He, Hongfang Yu, Jiawen Kang, Xiaoping Li, Zenglin
Xu, Dusit Niyato
- Abstract summary: FedGS is a hierarchical cloud-edge-end FL framework for 5G empowered industries.
Taking advantage of naturally clustered factory devices, FedGS uses a gradient-based binary permutation algorithm.
Experiments show that FedGS improves accuracy by 3.5% and reduces training rounds by 59% on average.
- Score: 57.67687126339891
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Nowadays, the industrial Internet of Things (IIoT) has played an integral
role in Industry 4.0 and produced massive amounts of data for industrial
intelligence. These data locate on decentralized devices in modern factories.
To protect the confidentiality of industrial data, federated learning (FL) was
introduced to collaboratively train shared machine learning models. However,
the local data collected by different devices skew in class distribution and
degrade industrial FL performance. This challenge has been widely studied at
the mobile edge, but they ignored the rapidly changing streaming data and
clustering nature of factory devices, and more seriously, they may threaten
data security. In this paper, we propose FedGS, which is a hierarchical
cloud-edge-end FL framework for 5G empowered industries, to improve industrial
FL performance on non-i.i.d. data. Taking advantage of naturally clustered
factory devices, FedGS uses a gradient-based binary permutation algorithm
(GBP-CS) to select a subset of devices within each factory and build
homogeneous super nodes participating in FL training. Then, we propose a
compound-step synchronization protocol to coordinate the training process
within and among these super nodes, which shows great robustness against data
heterogeneity. The proposed methods are time-efficient and can adapt to dynamic
environments, without exposing confidential industrial data in risky
manipulation. We prove that FedGS has better convergence performance than
FedAvg and give a relaxed condition under which FedGS is more
communication-efficient. Extensive experiments show that FedGS improves
accuracy by 3.5% and reduces training rounds by 59% on average, confirming its
superior effectiveness and efficiency on non-i.i.d. data.
Related papers
- Efficient Federated Intrusion Detection in 5G ecosystem using optimized BERT-based model [0.7100520098029439]
5G offers advanced services, supporting applications such as intelligent transportation, connected healthcare, and smart cities within the Internet of Things (IoT)
These advancements introduce significant security challenges, with increasingly sophisticated cyber-attacks.
This paper proposes a robust intrusion detection system (IDS) using federated learning and large language models (LLMs)
arXiv Detail & Related papers (2024-09-28T15:56:28Z) - A Quality-of-Service Compliance System using Federated Learning and
Optimistic Rollups [0.0]
A parallel trend is the rise of phones and tablets as primary computing devices for many people.
The powerful sensors present on these devices combined with the fact that they are mobile, mean they have access to data of an unprecedentedly diverse and private nature.
Models learned on such data hold the promise of greatly improving usability by powering more intelligent applications, but the sensitive nature of the data means there are risks and responsibilities to storing it in a centralized location.
We propose the use of Federated Learning (FL) so that specific data about services performed by clients do not leave the source machines.
arXiv Detail & Related papers (2023-11-14T20:02:37Z) - HFedMS: Heterogeneous Federated Learning with Memorable Data Semantics
in Industrial Metaverse [49.1501082763252]
This paper presents HFEDMS for incorporating practical FL into the emerging Industrial Metaverse.
It reduces data heterogeneity through dynamic grouping and training mode conversion.
Then, it compensates for the forgotten knowledge by fusing compressed historical data semantics.
Experiments have been conducted on the streamed non-i.i.d. FEMNIST dataset using 368 simulated devices.
arXiv Detail & Related papers (2022-11-07T04:33:24Z) - Online Data Selection for Federated Learning with Limited Storage [53.46789303416799]
Federated Learning (FL) has been proposed to achieve distributed machine learning among networked devices.
The impact of on-device storage on the performance of FL is still not explored.
In this work, we take the first step to consider the online data selection for FL with limited on-device storage.
arXiv Detail & Related papers (2022-09-01T03:27:33Z) - FedHiSyn: A Hierarchical Synchronous Federated Learning Framework for
Resource and Data Heterogeneity [56.82825745165945]
Federated Learning (FL) enables training a global model without sharing the decentralized raw data stored on multiple devices to protect data privacy.
We propose a hierarchical synchronous FL framework, i.e., FedHiSyn, to tackle the problems of straggler effects and outdated models.
We evaluate the proposed framework based on MNIST, EMNIST, CIFAR10 and CIFAR100 datasets and diverse heterogeneous settings of devices.
arXiv Detail & Related papers (2022-06-21T17:23:06Z) - Federated Dynamic Sparse Training: Computing Less, Communicating Less,
Yet Learning Better [88.28293442298015]
Federated learning (FL) enables distribution of machine learning workloads from the cloud to resource-limited edge devices.
We develop, implement, and experimentally validate a novel FL framework termed Federated Dynamic Sparse Training (FedDST)
FedDST is a dynamic process that extracts and trains sparse sub-networks from the target full network.
arXiv Detail & Related papers (2021-12-18T02:26:38Z) - Efficient Ring-topology Decentralized Federated Learning with Deep
Generative Models for Industrial Artificial Intelligent [13.982904025739606]
We propose a ring-topogy based decentralized federated learning scheme for Deep Generative Models (DGMs)
Our RDFL schemes provides communication efficiency and maintain training performance to boost DGMs in target IIoT tasks.
In addition, InterPlanetary File System(IPFS) is introduced to further improve communication efficiency and FL security.
arXiv Detail & Related papers (2021-04-15T08:09:54Z) - A Framework for Energy and Carbon Footprint Analysis of Distributed and
Federated Edge Learning [48.63610479916003]
This article breaks down and analyzes the main factors that influence the environmental footprint of distributed learning policies.
It models both vanilla and decentralized FL policies driven by consensus.
Results show that FL allows remarkable end-to-end energy savings (30%-40%) for wireless systems characterized by low bit/Joule efficiency.
arXiv Detail & Related papers (2021-03-18T16:04:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.