Dynamic Simplex: Balancing Safety and Performance in Autonomous Cyber
Physical Systems
- URL: http://arxiv.org/abs/2302.09750v1
- Date: Mon, 20 Feb 2023 04:00:53 GMT
- Title: Dynamic Simplex: Balancing Safety and Performance in Autonomous Cyber
Physical Systems
- Authors: Baiting Luo, Shreyas Ramakrishna, Ava Pettet, Christopher Kuhn, Gabor
Karsai, Ayan Mukhopadhyay
- Abstract summary: We propose a simplex strategy with an online controller switching logic that allows two-way switching.
We show that the proposed approach results in fewer collisions and higher performance than state-of-the-art alternatives.
- Score: 1.3309898919316483
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Learning Enabled Components (LEC) have greatly assisted cyber-physical
systems in achieving higher levels of autonomy. However, LEC's susceptibility
to dynamic and uncertain operating conditions is a critical challenge for the
safety of these systems. Redundant controller architectures have been widely
adopted for safety assurance in such contexts. These architectures augment LEC
"performant" controllers that are difficult to verify with "safety" controllers
and the decision logic to switch between them. While these architectures ensure
safety, we point out two limitations. First, they are trained offline to learn
a conservative policy of always selecting a controller that maintains the
system's safety, which limits the system's adaptability to dynamic and
non-stationary environments. Second, they do not support reverse switching from
the safety controller to the performant controller, even when the threat to
safety is no longer present. To address these limitations, we propose a dynamic
simplex strategy with an online controller switching logic that allows two-way
switching. We consider switching as a sequential decision-making problem and
model it as a semi-Markov decision process. We leverage a combination of a
myopic selector using surrogate models (for the forward switch) and a
non-myopic planner (for the reverse switch) to balance safety and performance.
We evaluate this approach using an autonomous vehicle case study in the CARLA
simulator using different driving conditions, locations, and component
failures. We show that the proposed approach results in fewer collisions and
higher performance than state-of-the-art alternatives.
Related papers
- Modular Control Architecture for Safe Marine Navigation: Reinforcement Learning and Predictive Safety Filters [0.0]
Reinforcement learning is increasingly used to adapt to complex scenarios, but standard frameworks ensuring safety and stability are lacking.
Predictive Safety Filters (PSF) offer a promising solution, ensuring constraint satisfaction in learning-based control without explicit constraint handling.
We apply this approach to marine navigation, combining RL with PSF on a simulated Cybership II model.
Results demonstrate the PSF's effectiveness in maintaining safety without hindering the RL agent's learning rate and performance, evaluated against a standard RL agent without PSF.
arXiv Detail & Related papers (2023-12-04T12:37:54Z) - Safe Deep Policy Adaptation [7.2747306035142225]
Policy adaptation based on reinforcement learning (RL) offers versatility and generalizability but presents safety and robustness challenges.
We propose SafeDPA, a novel RL and control framework that simultaneously tackles the problems of policy adaptation and safe reinforcement learning.
We provide theoretical safety guarantees of SafeDPA and show the robustness of SafeDPA against learning errors and extra perturbations.
arXiv Detail & Related papers (2023-10-08T00:32:59Z) - Safe Neural Control for Non-Affine Control Systems with Differentiable
Control Barrier Functions [58.19198103790931]
This paper addresses the problem of safety-critical control for non-affine control systems.
It has been shown that optimizing quadratic costs subject to state and control constraints can be sub-optimally reduced to a sequence of quadratic programs (QPs) by using Control Barrier Functions (CBFs)
We incorporate higher-order CBFs into neural ordinary differential equation-based learning models as differentiable CBFs to guarantee safety for non-affine control systems.
arXiv Detail & Related papers (2023-09-06T05:35:48Z) - Recursively Feasible Probabilistic Safe Online Learning with Control Barrier Functions [60.26921219698514]
We introduce a model-uncertainty-aware reformulation of CBF-based safety-critical controllers.
We then present the pointwise feasibility conditions of the resulting safety controller.
We use these conditions to devise an event-triggered online data collection strategy.
arXiv Detail & Related papers (2022-08-23T05:02:09Z) - A Safety-Critical Decision Making and Control Framework Combining
Machine Learning and Rule-based Algorithms [15.613795936398606]
Rule-based methods dominate in safety-critical systems. Yet, the latter cannot compete with the first ones in robustness to multiple requirements.
This paper proposes a decision making and control framework, which profits from advantages of both the rule- and machine-learning-based techniques.
arXiv Detail & Related papers (2022-01-30T13:58:51Z) - BarrierNet: A Safety-Guaranteed Layer for Neural Networks [50.86816322277293]
BarrierNet allows the safety constraints of a neural controller be adaptable to changing environments.
We evaluate them on a series of control problems such as traffic merging and robot navigations in 2D and 3D space.
arXiv Detail & Related papers (2021-11-22T15:38:11Z) - High-level Decisions from a Safe Maneuver Catalog with Reinforcement
Learning for Safe and Cooperative Automated Merging [5.732271870257913]
We propose an efficient RL-based decision-making pipeline for safe and cooperative automated driving in merging scenarios.
The proposed RL agent can efficiently identify cooperative drivers from their vehicle state history and generate interactive maneuvers.
arXiv Detail & Related papers (2021-07-15T15:49:53Z) - Non-stationary Online Learning with Memory and Non-stochastic Control [71.14503310914799]
We study the problem of Online Convex Optimization (OCO) with memory, which allows loss functions to depend on past decisions.
In this paper, we introduce dynamic policy regret as the performance measure to design algorithms robust to non-stationary environments.
We propose a novel algorithm for OCO with memory that provably enjoys an optimal dynamic policy regret in terms of time horizon, non-stationarity measure, and memory length.
arXiv Detail & Related papers (2021-02-07T09:45:15Z) - Enforcing robust control guarantees within neural network policies [76.00287474159973]
We propose a generic nonlinear control policy class, parameterized by neural networks, that enforces the same provable robustness criteria as robust control.
We demonstrate the power of this approach on several domains, improving in average-case performance over existing robust control methods and in worst-case stability over (non-robust) deep RL methods.
arXiv Detail & Related papers (2020-11-16T17:14:59Z) - Runtime Safety Assurance Using Reinforcement Learning [37.61747231296097]
This paper aims to design a meta-controller capable of identifying unsafe situations with high accuracy.
We frame the design of RTSA with the Markov decision process (MDP) and use reinforcement learning (RL) to solve it.
arXiv Detail & Related papers (2020-10-20T20:54:46Z) - Learning a Contact-Adaptive Controller for Robust, Efficient Legged
Locomotion [95.1825179206694]
We present a framework that synthesizes robust controllers for a quadruped robot.
A high-level controller learns to choose from a set of primitives in response to changes in the environment.
A low-level controller that utilizes an established control method to robustly execute the primitives.
arXiv Detail & Related papers (2020-09-21T16:49:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.