Multi-Stage Hybrid Federated Learning over Large-Scale D2D-Enabled Fog
Networks
- URL: http://arxiv.org/abs/2007.09511v5
- Date: Wed, 12 Jan 2022 22:43:01 GMT
- Title: Multi-Stage Hybrid Federated Learning over Large-Scale D2D-Enabled Fog
Networks
- Authors: Seyyedali Hosseinalipour, Sheikh Shams Azam, Christopher G. Brinton,
Nicolo Michelusi, Vaneet Aggarwal, David J. Love, and Huaiyu Dai
- Abstract summary: We develop a hybrid of intra- and inter-layer model learning that considers the network as a multi-layer cluster-based structure.
MH-FL considers the topology structures among the nodes in the clusters, including local networks formed via device-to-device (D2D) communications.
It orchestrates the devices at different network layers in a collaborative/cooperative manner to form local consensus on the model parameters.
- Score: 61.30171206892684
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Federated learning has generated significant interest, with nearly all works
focused on a "star" topology where nodes/devices are each connected to a
central server. We migrate away from this architecture and extend it through
the network dimension to the case where there are multiple layers of nodes
between the end devices and the server. Specifically, we develop multi-stage
hybrid federated learning (MH-FL), a hybrid of intra- and inter-layer model
learning that considers the network as a multi-layer cluster-based structure.
MH-FL considers the topology structures among the nodes in the clusters,
including local networks formed via device-to-device (D2D) communications, and
presumes a semi-decentralized architecture for federated learning. It
orchestrates the devices at different network layers in a
collaborative/cooperative manner (i.e., using D2D interactions) to form local
consensus on the model parameters and combines it with multi-stage parameter
relaying between layers of the tree-shaped hierarchy. We derive the upper bound
of convergence for MH-FL with respect to parameters of the network topology
(e.g., the spectral radius) and the learning algorithm (e.g., the number of D2D
rounds in different clusters). We obtain a set of policies for the D2D rounds
at different clusters to guarantee either a finite optimality gap or
convergence to the global optimum. We then develop a distributed control
algorithm for MH-FL to tune the D2D rounds in each cluster over time to meet
specific convergence criteria. Our experiments on real-world datasets verify
our analytical results and demonstrate the advantages of MH-FL in terms of
resource utilization metrics.
Related papers
- On Optimizing the Communication of Model Parallelism [74.15423270435949]
We study a novel and important communication pattern in large-scale model-parallel deep learning (DL)
In cross-mesh resharding, a sharded tensor needs to be sent from a source device mesh to a destination device mesh.
We propose two contributions to address cross-mesh resharding: an efficient broadcast-based communication system, and an "overlapping-friendly" pipeline schedule.
arXiv Detail & Related papers (2022-11-10T03:56:48Z) - Deep Image Clustering with Contrastive Learning and Multi-scale Graph
Convolutional Networks [58.868899595936476]
This paper presents a new deep clustering approach termed image clustering with contrastive learning and multi-scale graph convolutional networks (IcicleGCN)
Experiments on multiple image datasets demonstrate the superior clustering performance of IcicleGCN over the state-of-the-art.
arXiv Detail & Related papers (2022-07-14T19:16:56Z) - Deep Embedded Clustering with Distribution Consistency Preservation for
Attributed Networks [15.895606627146291]
In this study, we propose an end-to-end deep embedded clustering model for attributed networks.
It utilizes graph autoencoder and node attribute autoencoder to respectively learn node representations and cluster assignments.
The proposed model achieves significantly better or competitive performance compared with the state-of-the-art methods.
arXiv Detail & Related papers (2022-05-28T02:35:34Z) - Parallel Successive Learning for Dynamic Distributed Model Training over
Heterogeneous Wireless Networks [50.68446003616802]
Federated learning (FedL) has emerged as a popular technique for distributing model training over a set of wireless devices.
We develop parallel successive learning (PSL), which expands the FedL architecture along three dimensions.
Our analysis sheds light on the notion of cold vs. warmed up models, and model inertia in distributed machine learning.
arXiv Detail & Related papers (2022-02-07T05:11:01Z) - Federated Learning Beyond the Star: Local D2D Model Consensus with
Global Cluster Sampling [9.976132745670458]
Federated learning has emerged as a popular technique for distributing model training across the network edge.
We propose two timescale hybrid federated learning (TT-HF), which migrates to a more distributed topology via device-to-device (D2D) communications.
Experimental results demonstrate the improvements in convergence and utilization that can be obtained by TT-HF over state-of-the-art federated learning baselines.
arXiv Detail & Related papers (2021-09-07T21:48:17Z) - Two Timescale Hybrid Federated Learning with Cooperative D2D Local Model
Aggregations [10.702853653891902]
Federated learning has emerged as a popular technique for distributing machine learning (ML) model training across the wireless edge.
We propose two timescale hybrid federated learning (TT-HF), which is a hybrid between the device-to-server communication paradigm in federated learning and device-to-device (D2D) communications for model training.
arXiv Detail & Related papers (2021-03-18T18:58:45Z) - Multi-Level Local SGD for Heterogeneous Hierarchical Networks [11.699472346137739]
We propose Multi-Level Local SGD, a distributed gradient method for a learning, non- objective framework in a heterogeneous network.
We first provide a unified mathematical that describes the Multi-Level Local SGD algorithm.
We then present a theoretical analysis of the algorithm.
arXiv Detail & Related papers (2020-07-27T19:14:23Z) - From Federated to Fog Learning: Distributed Machine Learning over
Heterogeneous Wireless Networks [71.23327876898816]
Federated learning has emerged as a technique for training ML models at the network edge by leveraging processing capabilities across the nodes that collect the data.
We advocate a new learning paradigm called fog learning which will intelligently distribute ML model training across the continuum of nodes from edge devices to cloud servers.
arXiv Detail & Related papers (2020-06-07T05:11:18Z) - When Residual Learning Meets Dense Aggregation: Rethinking the
Aggregation of Deep Neural Networks [57.0502745301132]
We propose Micro-Dense Nets, a novel architecture with global residual learning and local micro-dense aggregations.
Our micro-dense block can be integrated with neural architecture search based models to boost their performance.
arXiv Detail & Related papers (2020-04-19T08:34:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.