Barrier Certified Safety Learning Control: When Sum-of-Square
Programming Meets Reinforcement Learning
- URL: http://arxiv.org/abs/2206.07915v1
- Date: Thu, 16 Jun 2022 04:38:50 GMT
- Title: Barrier Certified Safety Learning Control: When Sum-of-Square
Programming Meets Reinforcement Learning
- Authors: Hejun Huang, Zhenglong Li, Dongkun Han
- Abstract summary: This work adopts control barrier functions over reinforcement learning, and proposes a compensated algorithm to completely maintain safety.
Compared to quadratic programming based reinforcement learning methods, our sum-of-squares programming based reinforcement learning has shown its superiority.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Safety guarantee is essential in many engineering implementations.
Reinforcement learning provides a useful way to strengthen safety. However,
reinforcement learning algorithms cannot completely guarantee safety over
realistic operations. To address this issue, this work adopts control barrier
functions over reinforcement learning, and proposes a compensated algorithm to
completely maintain safety. Specifically, a sum-of-squares programming has been
exploited to search for the optimal controller, and tune the learning
hyperparameters simultaneously. Thus, the control actions are pledged to be
always within the safe region. The effectiveness of proposed method is
demonstrated via an inverted pendulum model. Compared to quadratic programming
based reinforcement learning methods, our sum-of-squares programming based
reinforcement learning has shown its superiority.
Related papers
- Safety through Permissibility: Shield Construction for Fast and Safe Reinforcement Learning [57.84059344739159]
"Shielding" is a popular technique to enforce safety inReinforcement Learning (RL)
We propose a new permissibility-based framework to deal with safety and shield construction.
arXiv Detail & Related papers (2024-05-29T18:00:21Z) - Learning Control Barrier Functions and their application in Reinforcement Learning: A Survey [11.180978323594822]
Reinforcement learning is a powerful technique for developing new robot behaviors.
It aims to incorporate safety considerations, enabling faster transfer to real robots and facilitating lifelong learning.
One promising approach within safe reinforcement learning is the use of control barrier functions.
arXiv Detail & Related papers (2024-04-22T22:52:14Z) - Reinforcement Learning with Ensemble Model Predictive Safety
Certification [2.658598582858331]
unsupervised exploration prevents the deployment of reinforcement learning algorithms on safety-critical tasks.
We propose a new algorithm that combines model-based deep reinforcement learning with tube-based model predictive control to correct the actions taken by a learning agent.
Our results show that we can achieve significantly fewer constraint violations than comparable reinforcement learning methods.
arXiv Detail & Related papers (2024-02-06T17:42:39Z) - Approximate Model-Based Shielding for Safe Reinforcement Learning [83.55437924143615]
We propose a principled look-ahead shielding algorithm for verifying the performance of learned RL policies.
Our algorithm differs from other shielding approaches in that it does not require prior knowledge of the safety-relevant dynamics of the system.
We demonstrate superior performance to other safety-aware approaches on a set of Atari games with state-dependent safety-labels.
arXiv Detail & Related papers (2023-07-27T15:19:45Z) - Adaptive Aggregation for Safety-Critical Control [3.1692938090731584]
We propose an adaptive aggregation framework for safety-critical control.
Our algorithm can achieve fewer safety violations while showing better data efficiency compared with several baselines.
arXiv Detail & Related papers (2023-02-07T16:53:33Z) - Evaluating Model-free Reinforcement Learning toward Safety-critical
Tasks [70.76757529955577]
This paper revisits prior work in this scope from the perspective of state-wise safe RL.
We propose Unrolling Safety Layer (USL), a joint method that combines safety optimization and safety projection.
To facilitate further research in this area, we reproduce related algorithms in a unified pipeline and incorporate them into SafeRL-Kit.
arXiv Detail & Related papers (2022-12-12T06:30:17Z) - Sample-efficient Safe Learning for Online Nonlinear Control with Control
Barrier Functions [35.9713619595494]
Reinforcement Learning and continuous nonlinear control have been successfully deployed in multiple domains of complicated sequential decision-making tasks.
Given the exploration nature of the learning process and the presence of model uncertainty, it is challenging to apply them to safety-critical control tasks.
We propose a emphprovably efficient episodic safe learning framework for online control tasks.
arXiv Detail & Related papers (2022-07-29T00:54:35Z) - Improving Safety in Deep Reinforcement Learning using Unsupervised
Action Planning [4.2955354157580325]
One of the key challenges to deep reinforcement learning (deep RL) is to ensure safety at both training and testing phases.
We propose a novel technique of unsupervised action planning to improve the safety of on-policy reinforcement learning algorithms.
Our results show that the proposed safety RL algorithm can achieve higher rewards compared with multiple baselines in both discrete and continuous control problems.
arXiv Detail & Related papers (2021-09-29T10:26:29Z) - Closing the Closed-Loop Distribution Shift in Safe Imitation Learning [80.05727171757454]
We treat safe optimization-based control strategies as experts in an imitation learning problem.
We train a learned policy that can be cheaply evaluated at run-time and that provably satisfies the same safety guarantees as the expert.
arXiv Detail & Related papers (2021-02-18T05:11:41Z) - Chance-Constrained Trajectory Optimization for Safe Exploration and
Learning of Nonlinear Systems [81.7983463275447]
Learning-based control algorithms require data collection with abundant supervision for training.
We present a new approach for optimal motion planning with safe exploration that integrates chance-constrained optimal control with dynamics learning and feedback control.
arXiv Detail & Related papers (2020-05-09T05:57:43Z) - Cautious Reinforcement Learning with Logical Constraints [78.96597639789279]
An adaptive safe padding forces Reinforcement Learning (RL) to synthesise optimal control policies while ensuring safety during the learning process.
Theoretical guarantees are available on the optimality of the synthesised policies and on the convergence of the learning algorithm.
arXiv Detail & Related papers (2020-02-26T00:01:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.