CorrFL: Correlation-Based Neural Network Architecture for Unavailability
Concerns in a Heterogeneous IoT Environment
- URL: http://arxiv.org/abs/2307.12149v1
- Date: Sat, 22 Jul 2023 19:23:06 GMT
- Title: CorrFL: Correlation-Based Neural Network Architecture for Unavailability
Concerns in a Heterogeneous IoT Environment
- Authors: Ibrahim Shaer, Abdallah Shami
- Abstract summary: The Federated Learning (FL) paradigm faces several challenges that limit its application in real-world environments.
These challenges include the local models' architecture heterogeneity and the unavailability of distributed Internet of Things (IoT) nodes due to connectivity problems.
This paper proposes the Correlation-based FL (CorrFL) approach influenced by the representational learning field to address this problem.
- Score: 3.9414768019101682
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The Federated Learning (FL) paradigm faces several challenges that limit its
application in real-world environments. These challenges include the local
models' architecture heterogeneity and the unavailability of distributed
Internet of Things (IoT) nodes due to connectivity problems. These factors
posit the question of "how can the available models fill the training gap of
the unavailable models?". This question is referred to as the "Oblique
Federated Learning" problem. This problem is encountered in the studied
environment that includes distributed IoT nodes responsible for predicting CO2
concentrations. This paper proposes the Correlation-based FL (CorrFL) approach
influenced by the representational learning field to address this problem.
CorrFL projects the various model weights to a common latent space to address
the model heterogeneity. Its loss function minimizes the reconstruction loss
when models are absent and maximizes the correlation between the generated
models. The latter factor is critical because of the intersection of the
feature spaces of the IoT devices. CorrFL is evaluated on a realistic use case,
involving the unavailability of one IoT device and heightened activity levels
that reflect occupancy. The generated CorrFL models for the unavailable IoT
device from the available ones trained on the new environment are compared
against models trained on different use cases, referred to as the benchmark
model. The evaluation criteria combine the mean absolute error (MAE) of
predictions and the impact of the amount of exchanged data on the prediction
performance improvement. Through a comprehensive experimental procedure, the
CorrFL model outperformed the benchmark model in every criterion.
Related papers
- SMILE: Zero-Shot Sparse Mixture of Low-Rank Experts Construction From Pre-Trained Foundation Models [85.67096251281191]
We present an innovative approach to model fusion called zero-shot Sparse MIxture of Low-rank Experts (SMILE) construction.
SMILE allows for the upscaling of source models into an MoE model without extra data or further training.
We conduct extensive experiments across diverse scenarios, such as image classification and text generation tasks, using full fine-tuning and LoRA fine-tuning.
arXiv Detail & Related papers (2024-08-19T17:32:15Z) - Stragglers-Aware Low-Latency Synchronous Federated Learning via Layer-Wise Model Updates [71.81037644563217]
Synchronous federated learning (FL) is a popular paradigm for collaborative edge learning.
As some of the devices may have limited computational resources and varying availability, FL latency is highly sensitive to stragglers.
We propose straggler-aware layer-wise federated learning (SALF) that leverages the optimization procedure of NNs via backpropagation to update the global model in a layer-wise fashion.
arXiv Detail & Related papers (2024-03-27T09:14:36Z) - Straggler-resilient Federated Learning: Tackling Computation
Heterogeneity with Layer-wise Partial Model Training in Mobile Edge Network [4.1813760301635705]
We propose Federated Partial Model Training (FedPMT), where devices with smaller computational capabilities work on partial models and contribute to the global model.
As such, all devices in FedPMT prioritize the most crucial parts of the global model.
Empirical results show that FedPMT significantly outperforms the existing benchmark FedDrop.
arXiv Detail & Related papers (2023-11-16T16:30:04Z) - A PAC-Bayesian Perspective on the Interpolating Information Criterion [54.548058449535155]
We show how a PAC-Bayes bound is obtained for a general class of models, characterizing factors which influence performance in the interpolating regime.
We quantify how the test error for overparameterized models achieving effectively zero training error depends on the quality of the implicit regularization imposed by e.g. the combination of model, parameter-initialization scheme.
arXiv Detail & Related papers (2023-11-13T01:48:08Z) - Over-the-Air Federated Learning and Optimization [52.5188988624998]
We focus on Federated learning (FL) via edge-the-air computation (AirComp)
We describe the convergence of AirComp-based FedAvg (AirFedAvg) algorithms under both convex and non- convex settings.
For different types of local updates that can be transmitted by edge devices (i.e., model, gradient, model difference), we reveal that transmitting in AirFedAvg may cause an aggregation error.
In addition, we consider more practical signal processing schemes to improve the communication efficiency and extend the convergence analysis to different forms of model aggregation error caused by these signal processing schemes.
arXiv Detail & Related papers (2023-10-16T05:49:28Z) - A Neuromorphic Architecture for Reinforcement Learning from Real-Valued
Observations [0.34410212782758043]
Reinforcement Learning (RL) provides a powerful framework for decision-making in complex environments.
This paper presents a novel Spiking Neural Network (SNN) architecture for solving RL problems with real-valued observations.
arXiv Detail & Related papers (2023-07-06T12:33:34Z) - A Distributed Computation Model Based on Federated Learning Integrates
Heterogeneous models and Consortium Blockchain for Solving Time-Varying
Problems [35.69540692050138]
We propose a Distributed Computation Model (DCM) based on the consortium blockchain network to improve the credibility of the overall model.
In the experiments, we verify the efficiency of DCM, where the results show that the proposed model outperforms many state-of-the-art models.
arXiv Detail & Related papers (2023-06-28T08:50:35Z) - Deep Equilibrium Models Meet Federated Learning [71.57324258813675]
This study explores the problem of Federated Learning (FL) by utilizing the Deep Equilibrium (DEQ) models instead of conventional deep learning networks.
We claim that incorporating DEQ models into the federated learning framework naturally addresses several open problems in FL.
To the best of our knowledge, this study is the first to establish a connection between DEQ models and federated learning.
arXiv Detail & Related papers (2023-05-29T22:51:40Z) - Towards Understanding and Mitigating Dimensional Collapse in Heterogeneous Federated Learning [112.69497636932955]
Federated learning aims to train models across different clients without the sharing of data for privacy considerations.
We study how data heterogeneity affects the representations of the globally aggregated models.
We propose sc FedDecorr, a novel method that can effectively mitigate dimensional collapse in federated learning.
arXiv Detail & Related papers (2022-10-01T09:04:17Z) - Decentralized Event-Triggered Federated Learning with Heterogeneous
Communication Thresholds [12.513477328344255]
We propose a novel methodology for distributed model aggregations via asynchronous, event-triggered consensus iterations over a network graph topology.
We demonstrate that our methodology achieves the globally optimal learning model under standard assumptions in distributed learning and graph consensus literature.
arXiv Detail & Related papers (2022-04-07T20:35:37Z) - FedHM: Efficient Federated Learning for Heterogeneous Models via
Low-rank Factorization [16.704006420306353]
A scalable federated learning framework should address heterogeneous clients equipped with different computation and communication capabilities.
This paper proposes FedHM, a novel federated model compression framework that distributes the heterogeneous low-rank models to clients and then aggregates them into a global full-rank model.
Our solution enables the training of heterogeneous local models with varying computational complexities and aggregates a single global model.
arXiv Detail & Related papers (2021-11-29T16:11:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.