Smart Scheduling based on Deep Reinforcement Learning for Cellular
Networks
- URL: http://arxiv.org/abs/2103.11542v1
- Date: Mon, 22 Mar 2021 02:09:16 GMT
- Title: Smart Scheduling based on Deep Reinforcement Learning for Cellular
Networks
- Authors: Jian Wang and Chen Xu and Rong Li and Yiqun Ge and Jun Wang
- Abstract summary: We propose a smart scheduling scheme based on deep reinforcement learning (DRL)
We provide implementation-friend designs, i.e., a scalable neural network design for the agent and a virtual environment training framework.
We show that the DRL-based smart scheduling outperforms the conventional scheduling method and can be adopted in practical systems.
- Score: 18.04856086228028
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: To improve the system performance towards the Shannon limit, advanced radio
resource management mechanisms play a fundamental role. In particular,
scheduling should receive much attention, because it allocates radio resources
among different users in terms of their channel conditions and QoS
requirements. The difficulties of scheduling algorithms are the tradeoffs need
to be made among multiple objectives, such as throughput, fairness and packet
drop rate. We propose a smart scheduling scheme based on deep reinforcement
learning (DRL). We not only verify the performance gain achieved, but also
provide implementation-friend designs, i.e., a scalable neural network design
for the agent and a virtual environment training framework. With the scalable
neural network design, the DRL agent can easily handle the cases when the
number of active users is time-varying without the need to redesign and retrain
the DRL agent. Training the DRL agent in a virtual environment offline first
and using it as the initial version in the practical usage helps to prevent the
system from suffering from performance and robustness degradation due to the
time-consuming training. Through both simulations and field tests, we show that
the DRL-based smart scheduling outperforms the conventional scheduling method
and can be adopted in practical systems.
Related papers
- Intent-Aware DRL-Based Uplink Dynamic Scheduler for 5G-NR [30.146175299047325]
We investigate the problem of supporting Industrial Internet of Things user equipment (IIoT UEs) with intent (i.e., requested quality of service (QoS)) and random traffic arrival.
A deep reinforcement learning (DRL) based centralized dynamic scheduler for time-frequency resources is proposed to learn how to schedule the available communication resources.
arXiv Detail & Related papers (2024-03-27T08:57:15Z) - Dynamic Scheduling for Federated Edge Learning with Streaming Data [56.91063444859008]
We consider a Federated Edge Learning (FEEL) system where training data are randomly generated over time at a set of distributed edge devices with long-term energy constraints.
Due to limited communication resources and latency requirements, only a subset of devices is scheduled for participating in the local training process in every iteration.
arXiv Detail & Related papers (2023-05-02T07:41:16Z) - HARL: Hierarchical Adaptive Reinforcement Learning Based Auto Scheduler
for Neural Networks [51.71682428015139]
We propose HARL, a reinforcement learning-based auto-scheduler for efficient tensor program exploration.
HarL improves the tensor operator performance by 22% and the search speed by 4.3x compared to the state-of-the-art auto-scheduler.
Inference performance and search speed are also significantly improved on end-to-end neural networks.
arXiv Detail & Related papers (2022-11-21T04:15:27Z) - FIRE: A Failure-Adaptive Reinforcement Learning Framework for Edge Computing Migrations [52.85536740465277]
FIRE is a framework that adapts to rare events by training a RL policy in an edge computing digital twin environment.
We propose ImRE, an importance sampling-based Q-learning algorithm, which samples rare events proportionally to their impact on the value function.
We show that FIRE reduces costs compared to vanilla RL and the greedy baseline in the event of failures.
arXiv Detail & Related papers (2022-09-28T19:49:39Z) - DL-DRL: A double-level deep reinforcement learning approach for
large-scale task scheduling of multi-UAV [65.07776277630228]
We propose a double-level deep reinforcement learning (DL-DRL) approach based on a divide and conquer framework (DCF)
Particularly, we design an encoder-decoder structured policy network in our upper-level DRL model to allocate the tasks to different UAVs.
We also exploit another attention based policy network in our lower-level DRL model to construct the route for each UAV, with the objective to maximize the number of executed tasks.
arXiv Detail & Related papers (2022-08-04T04:35:53Z) - Single-Shot Pruning for Offline Reinforcement Learning [47.886329599997474]
Deep Reinforcement Learning (RL) is a powerful framework for solving complex real-world problems.
One way to tackle this problem is to prune neural networks leaving only the necessary parameters.
We close the gap between RL and single-shot pruning techniques and present a general pruning approach to the Offline RL.
arXiv Detail & Related papers (2021-12-31T18:10:02Z) - MCDS: AI Augmented Workflow Scheduling in Mobile Edge Cloud Computing
Systems [12.215537834860699]
Recently proposed scheduling methods leverage the low response times of edge computing platforms to optimize application Quality of Service (QoS)
We propose MCDS: Monte Carlo Learning using Deep Surrogate Models to efficiently schedule workflow applications in mobile edge-cloud computing systems.
arXiv Detail & Related papers (2021-12-14T10:00:01Z) - Semantic-Aware Collaborative Deep Reinforcement Learning Over Wireless
Cellular Networks [82.02891936174221]
Collaborative deep reinforcement learning (CDRL) algorithms in which multiple agents can coordinate over a wireless network is a promising approach.
In this paper, a novel semantic-aware CDRL method is proposed to enable a group of untrained agents with semantically-linked DRL tasks to collaborate efficiently across a resource-constrained wireless cellular network.
arXiv Detail & Related papers (2021-11-23T18:24:47Z) - Deep-Reinforcement-Learning-Based Scheduling with Contiguous Resource
Allocation for Next-Generation Cellular Systems [4.227387975627387]
We propose a novel scheduling algorithm with contiguous frequency-domain resource allocation (FDRA) based on deep reinforcement learning (DRL)
The proposed DRL-based scheduling algorithm outperforms other representative baseline schemes while having lower online computational complexity.
arXiv Detail & Related papers (2020-10-11T05:41:40Z) - Knowledge-Assisted Deep Reinforcement Learning in 5G Scheduler Design:
From Theoretical Framework to Implementation [34.5517138843888]
We develop a knowledge-assisted deep reinforcement learning algorithm to design schedulers in 5G networks.
We show that a straightforward implementation of DDPG converges slowly, has a poor quality-of-service (QoS) performance, and cannot be implemented in real-world 5G systems.
arXiv Detail & Related papers (2020-09-17T14:52:12Z) - Stacked Auto Encoder Based Deep Reinforcement Learning for Online
Resource Scheduling in Large-Scale MEC Networks [44.40722828581203]
An online resource scheduling framework is proposed for minimizing the sum of weighted task latency for all the Internet of things (IoT) users.
A deep reinforcement learning (DRL) based solution is proposed, which includes the following components.
A preserved and prioritized experience replay (2p-ER) is introduced to assist the DRL to train the policy network and find the optimal offloading policy.
arXiv Detail & Related papers (2020-01-24T23:01:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.