Effective Multi-User Delay-Constrained Scheduling with Deep Recurrent
Reinforcement Learning
- URL: http://arxiv.org/abs/2208.14074v1
- Date: Tue, 30 Aug 2022 08:44:15 GMT
- Title: Effective Multi-User Delay-Constrained Scheduling with Deep Recurrent
Reinforcement Learning
- Authors: Pihe Hu, Ling Pan, Yu Chen, Zhixuan Fang, Longbo Huang
- Abstract summary: Multi-user delay constrained scheduling is important in many real-world applications including wireless communication, live streaming, and cloud computing.
We propose a deep reinforcement learning (DRL) algorithm, named Recurrent Softmax Delayed Deep Double Deterministic Policy Gradient ($mathttRSD4$)
$mathttRSD4$ guarantees resource and delay constraints by Lagrangian dual and delay-sensitive queues, respectively.
It also efficiently tackles partial observability with a memory mechanism enabled by the recurrent neural network (RNN) and introduces user-level decomposition and node-level
- Score: 28.35473469490186
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Multi-user delay constrained scheduling is important in many real-world
applications including wireless communication, live streaming, and cloud
computing. Yet, it poses a critical challenge since the scheduler needs to make
real-time decisions to guarantee the delay and resource constraints
simultaneously without prior information of system dynamics, which can be
time-varying and hard to estimate. Moreover, many practical scenarios suffer
from partial observability issues, e.g., due to sensing noise or hidden
correlation. To tackle these challenges, we propose a deep reinforcement
learning (DRL) algorithm, named Recurrent Softmax Delayed Deep Double
Deterministic Policy Gradient ($\mathtt{RSD4}$), which is a data-driven method
based on a Partially Observed Markov Decision Process (POMDP) formulation.
$\mathtt{RSD4}$ guarantees resource and delay constraints by Lagrangian dual
and delay-sensitive queues, respectively. It also efficiently tackles partial
observability with a memory mechanism enabled by the recurrent neural network
(RNN) and introduces user-level decomposition and node-level merging to ensure
scalability. Extensive experiments on simulated/real-world datasets demonstrate
that $\mathtt{RSD4}$ is robust to system dynamics and partially observable
environments, and achieves superior performances over existing DRL and
non-DRL-based methods.
Related papers
- Offline Critic-Guided Diffusion Policy for Multi-User Delay-Constrained Scheduling [29.431945795881976]
We propose a novel offline reinforcement learning-based algorithm, named underlineScheduling.
It learns efficient scheduling policies purely from pre-collected emphoffline data.
We show that SOCD is resilient to various system dynamics, including partially observable and large-scale environments.
arXiv Detail & Related papers (2025-01-22T15:13:21Z) - Latent feedback control of distributed systems in multiple scenarios through deep learning-based reduced order models [3.5161229331588095]
Continuous monitoring and real-time control of high-dimensional distributed systems are crucial in applications to ensure a desired physical behavior.
Traditional feedback control design that relies on full-order models fails to meet these requirements due to the delay in the control computation.
We propose a real-time closed-loop control strategy enhanced by nonlinear non-intrusive Deep Learning-based Reduced Order Models (DL-ROMs)
arXiv Detail & Related papers (2024-12-13T08:04:21Z) - Tractable Offline Learning of Regular Decision Processes [50.11277112628193]
This work studies offline Reinforcement Learning (RL) in a class of non-Markovian environments called Regular Decision Processes (RDPs)
Ins, the unknown dependency of future observations and rewards from the past interactions can be captured experimentally.
Many algorithms first reconstruct this unknown dependency using automata learning techniques.
arXiv Detail & Related papers (2024-09-04T14:26:58Z) - A Multi-Head Ensemble Multi-Task Learning Approach for Dynamical
Computation Offloading [62.34538208323411]
We propose a multi-head ensemble multi-task learning (MEMTL) approach with a shared backbone and multiple prediction heads (PHs)
MEMTL outperforms benchmark methods in both the inference accuracy and mean square error without requiring additional training data.
arXiv Detail & Related papers (2023-09-02T11:01:16Z) - Dynamic Scheduling for Federated Edge Learning with Streaming Data [56.91063444859008]
We consider a Federated Edge Learning (FEEL) system where training data are randomly generated over time at a set of distributed edge devices with long-term energy constraints.
Due to limited communication resources and latency requirements, only a subset of devices is scheduled for participating in the local training process in every iteration.
arXiv Detail & Related papers (2023-05-02T07:41:16Z) - Deep reinforcement learning under signal temporal logic constraints
using Lagrangian relaxation [0.0]
In general, a constraint may be imposed on the decision making.
We consider the optimal decision making problems with constraints to complete temporal high-level tasks.
We propose a two-phase constrained DRL algorithm using the Lagrangian relaxation method.
arXiv Detail & Related papers (2022-01-21T00:56:25Z) - Neural Stochastic Dual Dynamic Programming [99.80617899593526]
We introduce a trainable neural model that learns to map problem instances to a piece-wise linear value function.
$nu$-SDDP can significantly reduce problem solving cost without sacrificing solution quality.
arXiv Detail & Related papers (2021-12-01T22:55:23Z) - Dynamic Network-Assisted D2D-Aided Coded Distributed Learning [59.29409589861241]
We propose a novel device-to-device (D2D)-aided coded federated learning method (D2D-CFL) for load balancing across devices.
We derive an optimal compression rate for achieving minimum processing time and establish its connection with the convergence time.
Our proposed method is beneficial for real-time collaborative applications, where the users continuously generate training data.
arXiv Detail & Related papers (2021-11-26T18:44:59Z) - Short-Term Load Forecasting Using Time Pooling Deep Recurrent Neural
Network [0.0]
Integration of renewable energy sources and emerging loads like electric vehicles to smart grids brings more uncertainty to the distribution system management. Demand Side Management (DSM) is one of the approaches to reduce the uncertainty.
Some applications like Nonintrusive Load Monitoring (NILM) can support DSM, however they require accurate forecasting on high resolution data.
This is challenging when it comes to single loads like one residential household due to its high volatility.
arXiv Detail & Related papers (2021-09-26T05:20:48Z) - Revisiting State Augmentation methods for Reinforcement Learning with
Stochastic Delays [10.484851004093919]
This paper formally describes the notion of Markov Decision Processes (MDPs) with delays.
We show that delayed MDPs can be transformed into equivalent standard MDPs (without delays) with significantly simplified cost structure.
We employ this equivalence to derive a model-free Delay-Resolved RL framework and show that even a simple RL algorithm built upon this framework achieves near-optimal rewards in environments with delays in actions and observations.
arXiv Detail & Related papers (2021-08-17T10:45:55Z) - Combining Deep Learning and Optimization for Security-Constrained
Optimal Power Flow [94.24763814458686]
Security-constrained optimal power flow (SCOPF) is fundamental in power systems.
Modeling of APR within the SCOPF problem results in complex large-scale mixed-integer programs.
This paper proposes a novel approach that combines deep learning and robust optimization techniques.
arXiv Detail & Related papers (2020-07-14T12:38:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.