Fast Adaptive Task Offloading in Edge Computing based on Meta
Reinforcement Learning
- URL: http://arxiv.org/abs/2008.02033v5
- Date: Sat, 24 Oct 2020 10:04:20 GMT
- Title: Fast Adaptive Task Offloading in Edge Computing based on Meta
Reinforcement Learning
- Authors: Jin Wang, Jia Hu, Geyong Min, Albert Y. Zomaya, Nektarios Georgalas
- Abstract summary: Multi-access edge computing (MEC) aims to extend cloud service to the network edge to reduce network traffic and service latency.
A fundamental problem in MEC is how to efficiently offload heterogeneous tasks of mobile applications from user equipment (UE) to MEC hosts.
We propose a task offloading method based on meta reinforcement learning, which can adapt fast to new environments.
- Score: 44.81038225683222
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multi-access edge computing (MEC) aims to extend cloud service to the network
edge to reduce network traffic and service latency. A fundamental problem in
MEC is how to efficiently offload heterogeneous tasks of mobile applications
from user equipment (UE) to MEC hosts. Recently, many deep reinforcement
learning (DRL) based methods have been proposed to learn offloading policies
through interacting with the MEC environment that consists of UE, wireless
channels, and MEC hosts. However, these methods have weak adaptability to new
environments because they have low sample efficiency and need full retraining
to learn updated policies for new environments. To overcome this weakness, we
propose a task offloading method based on meta reinforcement learning, which
can adapt fast to new environments with a small number of gradient updates and
samples. We model mobile applications as Directed Acyclic Graphs (DAGs) and the
offloading policy by a custom sequence-to-sequence (seq2seq) neural network. To
efficiently train the seq2seq network, we propose a method that synergizes the
first order approximation and clipped surrogate objective. The experimental
results demonstrate that this new offloading method can reduce the latency by
up to 25% compared to three baselines while being able to adapt fast to new
environments.
Related papers
- A Multi-Head Ensemble Multi-Task Learning Approach for Dynamical
Computation Offloading [62.34538208323411]
We propose a multi-head ensemble multi-task learning (MEMTL) approach with a shared backbone and multiple prediction heads (PHs)
MEMTL outperforms benchmark methods in both the inference accuracy and mean square error without requiring additional training data.
arXiv Detail & Related papers (2023-09-02T11:01:16Z) - Resource Allocation in Multicore Elastic Optical Networks: A Deep
Reinforcement Learning Approach [47.187609203210705]
A new environment is developed compatible with OpenAI's Gym.
It processes the agent actions by considering the network state and physical-layer-related aspects.
The best-performing agent achieves a four-times decrease in blocking probability.
arXiv Detail & Related papers (2022-07-05T14:24:21Z) - Adapting to Dynamic LEO-B5G Systems: Meta-Critic Learning Based
Efficient Resource Scheduling [38.733584547351796]
We address two practical issues for an over-loaded LEO-terrestrial system.
The first challenge is how to efficiently schedule resources to serve the massive number of connected users.
The second challenge is how to make the algorithmic solution more resilient in adapting to dynamic wireless environments.
arXiv Detail & Related papers (2021-10-13T15:21:38Z) - Learning to Continuously Optimize Wireless Resource in a Dynamic
Environment: A Bilevel Optimization Perspective [52.497514255040514]
This work develops a new approach that enables data-driven methods to continuously learn and optimize resource allocation strategies in a dynamic environment.
We propose to build the notion of continual learning into wireless system design, so that the learning model can incrementally adapt to the new episodes.
Our design is based on a novel bilevel optimization formulation which ensures certain fairness" across different data samples.
arXiv Detail & Related papers (2021-05-03T07:23:39Z) - Efficient Feature Transformations for Discriminative and Generative
Continual Learning [98.10425163678082]
We propose a simple task-specific feature map transformation strategy for continual learning.
Theses provide powerful flexibility for learning new tasks, achieved with minimal parameters added to the base architecture.
We demonstrate the efficacy and efficiency of our method with an extensive set of experiments in discriminative (CIFAR-100 and ImageNet-1K) and generative sequences of tasks.
arXiv Detail & Related papers (2021-03-25T01:48:14Z) - Online Service Migration in Edge Computing with Incomplete Information:
A Deep Recurrent Actor-Critic Method [18.891775769665102]
Multi-access Edge Computing (MEC) is an emerging computing paradigm that extends cloud computing to the network edge.
Service migration needs to decide where to migrate user services for maintaining high Quality-of-Service (QoS)
We propose a new learning-driven method, namely Deep Recurrent ActorCritic based service Migration (DRACM), which is usercentric and can make effective online migration decisions.
arXiv Detail & Related papers (2020-12-16T00:16:24Z) - Learning to Continuously Optimize Wireless Resource In Episodically
Dynamic Environment [55.91291559442884]
This work develops a methodology that enables data-driven methods to continuously learn and optimize in a dynamic environment.
We propose to build the notion of continual learning into the modeling process of learning wireless systems.
Our design is based on a novel min-max formulation which ensures certain fairness" across different data samples.
arXiv Detail & Related papers (2020-11-16T08:24:34Z) - Deep Reinforcement Learning for Adaptive Network Slicing in 5G for
Intelligent Vehicular Systems and Smart Cities [19.723551683930776]
We develop a network slicing model based on a cluster of fog nodes (FNs) coordinated with an edge controller (EC)
For each service request in a cluster, the EC decides which FN to execute the task, locally serve the request at the edge, or to reject the task and refer it to the cloud.
We propose a deep reinforcement learning (DRL) solution to adaptively learn the optimal slicing policy.
arXiv Detail & Related papers (2020-10-19T23:30:08Z) - Meta Reinforcement Learning with Autonomous Inference of Subtask
Dependencies [57.27944046925876]
We propose and address a novel few-shot RL problem, where a task is characterized by a subtask graph.
Instead of directly learning a meta-policy, we develop a Meta-learner with Subtask Graph Inference.
Our experiment results on two grid-world domains and StarCraft II environments show that the proposed method is able to accurately infer the latent task parameter.
arXiv Detail & Related papers (2020-01-01T17:34:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.