Knowledge Distillation for Mobile Edge Computation Offloading
- URL: http://arxiv.org/abs/2004.04366v1
- Date: Thu, 9 Apr 2020 04:58:46 GMT
- Title: Knowledge Distillation for Mobile Edge Computation Offloading
- Authors: Haowei Chen, Liekang Zeng, Shuai Yu, and Xu Chen
- Abstract summary: We propose an edge computation offloading framework based on Deep Imitation Learning (DIL) and Knowledge Distillation (KD)
Our model has the shortest inference delay among all policies.
- Score: 14.417463848473494
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Edge computation offloading allows mobile end devices to put execution of
compute-intensive task on the edge servers. End devices can decide whether
offload the tasks to edge servers, cloud servers or execute locally according
to current network condition and devices' profile in an online manner. In this
article, we propose an edge computation offloading framework based on Deep
Imitation Learning (DIL) and Knowledge Distillation (KD), which assists end
devices to quickly make fine-grained decisions to optimize the delay of
computation tasks online. We formalize computation offloading problem into a
multi-label classification problem. Training samples for our DIL model are
generated in an offline manner. After model is trained, we leverage knowledge
distillation to obtain a lightweight DIL model, by which we further reduce the
model's inference delay. Numerical experiment shows that the offloading
decisions made by our model outperforms those made by other related policies in
latency metric. Also, our model has the shortest inference delay among all
policies.
Related papers
- Efficient Asynchronous Federated Learning with Sparsification and
Quantization [55.6801207905772]
Federated Learning (FL) is attracting more and more attention to collaboratively train a machine learning model without transferring raw data.
FL generally exploits a parameter server and a large number of edge devices during the whole process of the model training.
We propose TEASQ-Fed to exploit edge devices to asynchronously participate in the training process by actively applying for tasks.
arXiv Detail & Related papers (2023-12-23T07:47:07Z) - Fractional Deep Reinforcement Learning for Age-Minimal Mobile Edge
Computing [11.403989519949173]
This work focuses on the timeliness of computational-intensive updates, measured by Age-ofInformation (AoI)
We study how to jointly optimize the task updating and offloading policies for AoI with fractional form.
Experimental results show that our proposed algorithms reduce the average AoI by up to 57.6% compared with several non-fractional benchmarks.
arXiv Detail & Related papers (2023-12-16T11:13:40Z) - Age-Based Scheduling for Mobile Edge Computing: A Deep Reinforcement
Learning Approach [58.911515417156174]
We propose a new definition of Age of Information (AoI) and, based on the redefined AoI, we formulate an online AoI problem for MEC systems.
We introduce Post-Decision States (PDSs) to exploit the partial knowledge of the system's dynamics.
We also combine PDSs with deep RL to further improve the algorithm's applicability, scalability, and robustness.
arXiv Detail & Related papers (2023-12-01T01:30:49Z) - Scheduling Inference Workloads on Distributed Edge Clusters with
Reinforcement Learning [11.007816552466952]
This paper focuses on the problem of scheduling inference queries on Deep Neural Networks in edge networks at short timescales.
By means of simulations, we analyze several policies in the realistic network settings and workloads of a large ISP.
We design ASET, a Reinforcement Learning based scheduling algorithm able to adapt its decisions according to the system conditions.
arXiv Detail & Related papers (2023-01-31T13:23:34Z) - Graph Reinforcement Learning-based CNN Inference Offloading in Dynamic
Edge Computing [93.67044879636093]
This paper addresses the computational offloading of CNN inference in dynamic multi-access edge computing (MEC) networks.
We propose a graph reinforcement learning-based early-exit mechanism (GRLE) which outperforms the state-of-the-art work.
The experimental results show that GRLE achieves the average accuracy up to 3.41x over graph reinforcement learning (GRL) and 1.45x over DROOE.
arXiv Detail & Related papers (2022-10-24T07:17:20Z) - Online Learning for Orchestration of Inference in Multi-User
End-Edge-Cloud Networks [3.6076391721440633]
Collaborative end-edge-cloud computing for deep learning provides a range of performance and efficiency.
We propose a reinforcement-learning-based computation offloading solution that learns optimal offloading policy.
Our solution provides 35% speedup in the average response time compared to the state-of-the-art with less than 0.9% accuracy reduction.
arXiv Detail & Related papers (2022-02-21T21:41:29Z) - Offloading Algorithms for Maximizing Inference Accuracy on Edge Device
Under a Time Constraint [15.038891477389535]
We propose an approximation algorithm AMR2, and prove that it results in a makespan at most 2T, and achieves a total accuracy that is lower by a small constant from optimal total accuracy.
As proof of concept, we implemented AMR2 on a Raspberry Pi, equipped with MobileNet, and is connected to a server equipped with ResNet, and studied the total accuracy and makespan performance of AMR2 for image classification application.
arXiv Detail & Related papers (2021-12-21T18:21:24Z) - Computational Intelligence and Deep Learning for Next-Generation
Edge-Enabled Industrial IoT [51.68933585002123]
We investigate how to deploy computational intelligence and deep learning (DL) in edge-enabled industrial IoT networks.
In this paper, we propose a novel multi-exit-based federated edge learning (ME-FEEL) framework.
In particular, the proposed ME-FEEL can achieve an accuracy gain up to 32.7% in the industrial IoT networks with the severely limited resources.
arXiv Detail & Related papers (2021-10-28T08:14:57Z) - Real-Time Edge Classification: Optimal Offloading under Token Bucket
Constraints [13.583977689847433]
We introduce a Markov Decision Process-based framework to make offload decisions under strict latency constraints.
We also propose approaches to allow multiple devices connected to the same access switch to share their bursting allocation.
We evaluate and analyze the policies derived using our framework on the standard ImageNet image classification benchmark.
arXiv Detail & Related papers (2020-10-26T17:25:29Z) - Adaptive Subcarrier, Parameter, and Power Allocation for Partitioned
Edge Learning Over Broadband Channels [69.18343801164741]
partitioned edge learning (PARTEL) implements parameter-server training, a well known distributed learning method, in wireless network.
We consider the case of deep neural network (DNN) models which can be trained using PARTEL by introducing some auxiliary variables.
arXiv Detail & Related papers (2020-10-08T15:27:50Z) - Joint Parameter-and-Bandwidth Allocation for Improving the Efficiency of
Partitioned Edge Learning [73.82875010696849]
Machine learning algorithms are deployed at the network edge for training artificial intelligence (AI) models.
This paper focuses on the novel joint design of parameter (computation load) allocation and bandwidth allocation.
arXiv Detail & Related papers (2020-03-10T05:52:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.