Safe RAN control: A Symbolic Reinforcement Learning Approach
- URL: http://arxiv.org/abs/2106.01977v1
- Date: Thu, 3 Jun 2021 16:45:40 GMT
- Title: Safe RAN control: A Symbolic Reinforcement Learning Approach
- Authors: Alexandros Nikou, Anusha Mujumdar, Marin Orlic, Aneta Vulgarakis
Feljan
- Abstract summary: We present a Symbolic Reinforcement Learning (SRL) based architecture for safety control of Radio Access Network (RAN) applications.
We provide a purely automated procedure in which a user can specify high-level logical safety specifications for a given cellular network topology.
We introduce a user interface (UI) developed to help a user set intent specifications to the system, and inspect the difference in agent proposed actions.
- Score: 62.997667081978825
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we present a Symbolic Reinforcement Learning (SRL) based
architecture for safety control of Radio Access Network (RAN) applications. In
particular, we provide a purely automated procedure in which a user can specify
high-level logical safety specifications for a given cellular network topology
in order for the latter to execute optimal safe performance which is measured
through certain Key Performance Indicators (KPIs). The network consists of a
set of fixed Base Stations (BS) which are equipped with antennas, which one can
control by adjusting their vertical tilt angle. The aforementioned process is
called Remote Electrical Tilt (RET) optimization. Recent research has focused
on performing this RET optimization by employing Reinforcement Learning (RL)
strategies due to the fact that they have self-learning capabilities to adapt
in uncertain environments. The term safety refers to particular constraints
bounds of the network KPIs in order to guarantee that when the algorithms are
deployed in a live network, the performance is maintained. In our proposed
architecture the safety is ensured through model-checking techniques over
combined discrete system models (automata) that are abstracted through the
learning process. We introduce a user interface (UI) developed to help a user
set intent specifications to the system, and inspect the difference in agent
proposed actions, and those that are allowed and blocked according to the
safety specification.
Related papers
- Online Safety Property Collection and Refinement for Safe Deep
Reinforcement Learning in Mapless Navigation [79.89605349842569]
We introduce the Collection and Refinement of Online Properties (CROP) framework to design properties at training time.
CROP employs a cost signal to identify unsafe interactions and use them to shape safety properties.
We evaluate our approach in several robotic mapless navigation tasks and demonstrate that the violation metric computed with CROP allows higher returns and lower violations over previous Safe DRL approaches.
arXiv Detail & Related papers (2023-02-13T21:19:36Z) - Evaluating Model-free Reinforcement Learning toward Safety-critical
Tasks [70.76757529955577]
This paper revisits prior work in this scope from the perspective of state-wise safe RL.
We propose Unrolling Safety Layer (USL), a joint method that combines safety optimization and safety projection.
To facilitate further research in this area, we reproduce related algorithms in a unified pipeline and incorporate them into SafeRL-Kit.
arXiv Detail & Related papers (2022-12-12T06:30:17Z) - Recursively Feasible Probabilistic Safe Online Learning with Control Barrier Functions [60.26921219698514]
We introduce a model-uncertainty-aware reformulation of CBF-based safety-critical controllers.
We then present the pointwise feasibility conditions of the resulting safety controller.
We use these conditions to devise an event-triggered online data collection strategy.
arXiv Detail & Related papers (2022-08-23T05:02:09Z) - Safe-Critical Modular Deep Reinforcement Learning with Temporal Logic
through Gaussian Processes and Control Barrier Functions [3.5897534810405403]
Reinforcement learning (RL) is a promising approach and has limited success towards real-world applications.
In this paper, we propose a learning-based control framework consisting of several aspects.
We show such an ECBF-based modular deep RL algorithm achieves near-perfect success rates and guard safety with a high probability.
arXiv Detail & Related papers (2021-09-07T00:51:12Z) - Symbolic Reinforcement Learning for Safe RAN Control [62.997667081978825]
We show a Symbolic Reinforcement Learning (SRL) architecture for safe control in Radio Access Network (RAN) applications.
In our tool, a user can select a high-level safety specifications expressed in Linear Temporal Logic (LTL) to shield an RL agent running in a given cellular network.
We demonstrate the user interface (UI) helping the user set intent specifications to the architecture and inspect the difference in allowed and blocked actions.
arXiv Detail & Related papers (2021-03-11T10:56:49Z) - A Safe Reinforcement Learning Architecture for Antenna Tilt Optimisation [2.099922236065961]
Safe interaction with the environment is one of the most challenging aspects of Reinforcement Learning (RL) when applied to real-world problems.
Remote Electrical Tilt (RET) optimisation is a safety-critical application in which exploratory modifications of antenna tilt angles of base stations can cause significant performance degradation in the network.
We propose a modular Safe Reinforcement Learning architecture which is then used to address the RET optimisation in cellular networks.
arXiv Detail & Related papers (2020-12-02T16:07:55Z) - Remote Electrical Tilt Optimization via Safe Reinforcement Learning [1.2599533416395765]
Remote Electrical Tilt (RET) optimization is an efficient method for adjusting the vertical tilt angle of Base Stations (BSs) antennas in order to optimize Key Performance Indicators (KPIs) of the network.
In this work, we model the RET optimization problem in the Safe Reinforcement Learning (SRL) framework with the goal of learning a tilt control strategy.
Our experiments show that the proposed approach is able to learn a safe and improved tilt update policy, providing a higher degree of reliability and potential for real-world network deployment.
arXiv Detail & Related papers (2020-10-12T16:46:40Z) - Certified Reinforcement Learning with Logic Guidance [78.2286146954051]
We propose a model-free RL algorithm that enables the use of Linear Temporal Logic (LTL) to formulate a goal for unknown continuous-state/action Markov Decision Processes (MDPs)
The algorithm is guaranteed to synthesise a control policy whose traces satisfy the specification with maximal probability.
arXiv Detail & Related papers (2019-02-02T20:09:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.