An Efficient Split Fine-tuning Framework for Edge and Cloud
Collaborative Learning
- URL: http://arxiv.org/abs/2211.16703v1
- Date: Wed, 30 Nov 2022 02:55:21 GMT
- Title: An Efficient Split Fine-tuning Framework for Edge and Cloud
Collaborative Learning
- Authors: Shaohuai Shi, Qing Yang, Yang Xiang, Shuhan Qi, Xuan Wang
- Abstract summary: We design an efficient split fine-tuning framework for edge and cloud collaborative learning.
We compress the intermediate output of a neural network to reduce the communication volume between the edge device and the cloud server.
Our framework can reduce the communication traffic by 96 times with little impact on the model accuracy.
- Score: 20.118073642453034
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: To enable the pre-trained models to be fine-tuned with local data on edge
devices without sharing data with the cloud, we design an efficient split
fine-tuning (SFT) framework for edge and cloud collaborative learning. We
propose three novel techniques in this framework. First, we propose a matrix
decomposition-based method to compress the intermediate output of a neural
network to reduce the communication volume between the edge device and the
cloud server. Second, we eliminate particular links in the model without
affecting the convergence performance in fine-tuning. Third, we implement our
system atop PyTorch to allow users to easily extend their existing training
scripts to enjoy the efficient edge and cloud collaborative learning.
Experiments results on 9 NLP datasets show that our framework can reduce the
communication traffic by 96 times with little impact on the model accuracy.
Related papers
- P2P-Bridge: Diffusion Bridges for 3D Point Cloud Denoising [81.92854168911704]
We tackle the task of point cloud denoising through a novel framework that adapts Diffusion Schr"odinger bridges to points clouds.
Experiments on object datasets show that P2P-Bridge achieves significant improvements over existing methods.
arXiv Detail & Related papers (2024-08-29T08:00:07Z) - ECLM: Efficient Edge-Cloud Collaborative Learning with Continuous
Environment Adaptation [47.35179593006409]
We propose ECLM, an edge-cloud collaborative learning framework for rapid model adaptation for dynamic edge environments.
We show that ECLM significantly improves model performance (e.g., 18.89% accuracy increase) and resource efficiency (e.g. 7.12x communication cost reduction) in adapting models to dynamic edge environments.
arXiv Detail & Related papers (2023-11-18T14:10:09Z) - Shoggoth: Towards Efficient Edge-Cloud Collaborative Real-Time Video
Inference via Adaptive Online Learning [33.16911236522438]
Shoggoth is an efficient edge-cloud collaborative architecture for boosting inference performance on real-time video of changing scenes.
Online knowledge distillation improves the accuracy of models suffering from data drift and offloads the labeling process to the cloud.
At the edge, we design adaptive training using small batches to adapt models under limited computing power.
arXiv Detail & Related papers (2023-06-27T09:39:42Z) - Data Augmentation-free Unsupervised Learning for 3D Point Cloud
Understanding [61.30276576646909]
We propose an augmentation-free unsupervised approach for point clouds to learn transferable point-level features via soft clustering, named SoftClu.
We exploit the affiliation of points to their clusters as a proxy to enable self-training through a pseudo-label prediction task.
arXiv Detail & Related papers (2022-10-06T10:18:16Z) - Receptive Field-based Segmentation for Distributed CNN Inference
Acceleration in Collaborative Edge Computing [93.67044879636093]
We study inference acceleration using distributed convolutional neural networks (CNNs) in collaborative edge computing network.
We propose a novel collaborative edge computing using fused-layer parallelization to partition a CNN model into multiple blocks of convolutional layers.
arXiv Detail & Related papers (2022-07-22T18:38:11Z) - Semi-Decentralized Federated Edge Learning with Data and Device
Heterogeneity [6.341508488542275]
Federated edge learning (FEEL) has attracted much attention as a privacy-preserving paradigm to effectively incorporate the distributed data at the network edge for training deep learning models.
In this paper, we investigate a novel framework of FEEL, namely semi-decentralized federated edge learning (SD-FEEL), where multiple edge servers are employed to collectively coordinate a large number of client nodes.
By exploiting the low-latency communication among edge servers for efficient model sharing, SD-FEEL can incorporate more training data, while enjoying much lower latency compared with conventional federated learning.
arXiv Detail & Related papers (2021-12-20T03:06:08Z) - ProgFed: Effective, Communication, and Computation Efficient Federated Learning by Progressive Training [65.68511423300812]
We propose ProgFed, a progressive training framework for efficient and effective federated learning.
ProgFed inherently reduces computation and two-way communication costs while maintaining the strong performance of the final models.
Our results show that ProgFed converges at the same rate as standard training on full models.
arXiv Detail & Related papers (2021-10-11T14:45:00Z) - Complexity-aware Adaptive Training and Inference for Edge-Cloud
Distributed AI Systems [9.273593723275544]
IoT and machine learning applications create large amounts of data that require real-time processing.
We propose a distributed AI system to exploit both the edge and the cloud for training and inference.
arXiv Detail & Related papers (2021-09-14T05:03:54Z) - Hierarchical Quantized Federated Learning: Convergence Analysis and
System Design [7.481427303081613]
Federated learning is a collaborative machine to train deep neural networks without clients' private data.
Previous works assume one central parameter either at the cloud or at the edge.
This paper exploits the advantages of both cloud servers and considers Hierarchical Quantized Federated Learning system.
arXiv Detail & Related papers (2021-03-26T05:48:36Z) - Joint Parameter-and-Bandwidth Allocation for Improving the Efficiency of
Partitioned Edge Learning [73.82875010696849]
Machine learning algorithms are deployed at the network edge for training artificial intelligence (AI) models.
This paper focuses on the novel joint design of parameter (computation load) allocation and bandwidth allocation.
arXiv Detail & Related papers (2020-03-10T05:52:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.