Structure-Enhanced DRL for Optimal Transmission Scheduling
- URL: http://arxiv.org/abs/2212.12704v1
- Date: Sat, 24 Dec 2022 10:18:38 GMT
- Title: Structure-Enhanced DRL for Optimal Transmission Scheduling
- Authors: Jiazheng Chen, Wanchun Liu, Daniel E. Quevedo, Saeed R. Khosravirad,
Yonghui Li, and Branka Vucetic
- Abstract summary: This paper focuses on the transmission scheduling problem of a remote estimation system.
We develop a structure-enhanced deep reinforcement learning framework for optimal scheduling of the system.
In particular, we propose a structure-enhanced action selection method, which tends to select actions that obey the policy structure.
- Score: 43.801422320012286
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Remote state estimation of large-scale distributed dynamic processes plays an
important role in Industry 4.0 applications. In this paper, we focus on the
transmission scheduling problem of a remote estimation system. First, we derive
some structural properties of the optimal sensor scheduling policy over fading
channels. Then, building on these theoretical guidelines, we develop a
structure-enhanced deep reinforcement learning (DRL) framework for optimal
scheduling of the system to achieve the minimum overall estimation mean-square
error (MSE). In particular, we propose a structure-enhanced action selection
method, which tends to select actions that obey the policy structure. This
explores the action space more effectively and enhances the learning efficiency
of DRL agents. Furthermore, we introduce a structure-enhanced loss function to
add penalties to actions that do not follow the policy structure. The new loss
function guides the DRL to converge to the optimal policy structure quickly.
Our numerical experiments illustrate that the proposed structure-enhanced DRL
algorithms can save the training time by 50% and reduce the remote estimation
MSE by 10% to 25% when compared to benchmark DRL algorithms. In addition, we
show that the derived structural properties exist in a wide range of dynamic
scheduling problems that go beyond remote state estimation.
Related papers
- Truncating Trajectories in Monte Carlo Policy Evaluation: an Adaptive Approach [51.76826149868971]
Policy evaluation via Monte Carlo simulation is at the core of many MC Reinforcement Learning (RL) algorithms.
We propose as a quality index a surrogate of the mean squared error of a return estimator that uses trajectories of different lengths.
We present an adaptive algorithm called Robust and Iterative Data collection strategy Optimization (RIDO)
arXiv Detail & Related papers (2024-10-17T11:47:56Z) - Optimization Theory Based Deep Reinforcement Learning for Resource
Allocation in Ultra-Reliable Wireless Networked Control Systems [10.177917426690701]
This paper introduces a novel optimization theory based deep reinforcement learning (DRL) framework for the joint design of controller and communication systems.
The objective of minimum power consumption is targeted while satisfying the schedulability and rate constraints of the communication system.
arXiv Detail & Related papers (2023-11-28T15:49:29Z) - Structure-Enhanced Deep Reinforcement Learning for Optimal Transmission
Scheduling [47.29474858956844]
We develop a structure-enhanced deep reinforcement learning framework for optimal scheduling of a multi-sensor remote estimation system.
In particular, we propose a structure-enhanced action selection method, which tends to select actions that obey the policy structure.
Our numerical results show that the proposed structure-enhanced DRL algorithms can save the training time by 50% and reduce the remote estimation MSE by 10% to 25%.
arXiv Detail & Related papers (2022-11-20T00:13:35Z) - CACTO: Continuous Actor-Critic with Trajectory Optimization -- Towards
global optimality [5.0915256711576475]
This paper presents a novel algorithm for the continuous control of dynamical systems that combines Trayy (TO) and Reinforcement Learning (RL) in a single trajectory.
arXiv Detail & Related papers (2022-11-12T10:16:35Z) - Towards Deployment-Efficient Reinforcement Learning: Lower Bound and
Optimality [141.89413461337324]
Deployment efficiency is an important criterion for many real-world applications of reinforcement learning (RL)
We propose a theoretical formulation for deployment-efficient RL (DE-RL) from an "optimization with constraints" perspective.
arXiv Detail & Related papers (2022-02-14T01:31:46Z) - RL-Controller: a reinforcement learning framework for active structural
control [0.0]
We present a novel RL-based approach for designing active controllers by introducing RL-Controller, a flexible and scalable simulation environment.
We show that the proposed framework is easily trainable for a five story benchmark building with 65% reductions on average in inter story drifts.
In a comparative study with LQG active control method, we demonstrate that the proposed model-free algorithm learns more optimal actuator forcing strategies.
arXiv Detail & Related papers (2021-03-13T04:42:13Z) - Optimization-driven Deep Reinforcement Learning for Robust Beamforming
in IRS-assisted Wireless Communications [54.610318402371185]
Intelligent reflecting surface (IRS) is a promising technology to assist downlink information transmissions from a multi-antenna access point (AP) to a receiver.
We minimize the AP's transmit power by a joint optimization of the AP's active beamforming and the IRS's passive beamforming.
We propose a deep reinforcement learning (DRL) approach that can adapt the beamforming strategies from past experiences.
arXiv Detail & Related papers (2020-05-25T01:42:55Z) - Guided Constrained Policy Optimization for Dynamic Quadrupedal Robot
Locomotion [78.46388769788405]
We introduce guided constrained policy optimization (GCPO), an RL framework based upon our implementation of constrained policy optimization (CPPO)
We show that guided constrained RL offers faster convergence close to the desired optimum resulting in an optimal, yet physically feasible, robotic control behavior without the need for precise reward function tuning.
arXiv Detail & Related papers (2020-02-22T10:15:53Z) - Stacked Auto Encoder Based Deep Reinforcement Learning for Online
Resource Scheduling in Large-Scale MEC Networks [44.40722828581203]
An online resource scheduling framework is proposed for minimizing the sum of weighted task latency for all the Internet of things (IoT) users.
A deep reinforcement learning (DRL) based solution is proposed, which includes the following components.
A preserved and prioritized experience replay (2p-ER) is introduced to assist the DRL to train the policy network and find the optimal offloading policy.
arXiv Detail & Related papers (2020-01-24T23:01:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.