A Safety-Critical Decision Making and Control Framework Combining
Machine Learning and Rule-based Algorithms
- URL: http://arxiv.org/abs/2201.12819v1
- Date: Sun, 30 Jan 2022 13:58:51 GMT
- Title: A Safety-Critical Decision Making and Control Framework Combining
Machine Learning and Rule-based Algorithms
- Authors: Andrei Aksjonov and Ville Kyrki
- Abstract summary: Rule-based methods dominate in safety-critical systems. Yet, the latter cannot compete with the first ones in robustness to multiple requirements.
This paper proposes a decision making and control framework, which profits from advantages of both the rule- and machine-learning-based techniques.
- Score: 15.613795936398606
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: While artificial-intelligence-based methods suffer from lack of transparency,
rule-based methods dominate in safety-critical systems. Yet, the latter cannot
compete with the first ones in robustness to multiple requirements, for
instance, simultaneously addressing safety, comfort, and efficiency. Hence, to
benefit from both methods they must be joined in a single system. This paper
proposes a decision making and control framework, which profits from advantages
of both the rule- and machine-learning-based techniques while compensating for
their disadvantages. The proposed method embodies two controllers operating in
parallel, called Safety and Learned. A rule-based switching logic selects one
of the actions transmitted from both controllers. The Safety controller is
prioritized every time, when the Learned one does not meet the safety
constraint, and also directly participates in the safe Learned controller
training. Decision making and control in autonomous driving is chosen as the
system case study, where an autonomous vehicle learns a multi-task policy to
safely cross an unprotected intersection. Multiple requirements (i.e., safety,
efficiency, and comfort) are set for vehicle operation. A numerical simulation
is performed for the proposed framework validation, where its ability to
satisfy the requirements and robustness to changing environment is successfully
demonstrated.
Related papers
- Dynamic Simplex: Balancing Safety and Performance in Autonomous Cyber
Physical Systems [1.3309898919316483]
We propose a simplex strategy with an online controller switching logic that allows two-way switching.
We show that the proposed approach results in fewer collisions and higher performance than state-of-the-art alternatives.
arXiv Detail & Related papers (2023-02-20T04:00:53Z) - Safety Correction from Baseline: Towards the Risk-aware Policy in
Robotics via Dual-agent Reinforcement Learning [64.11013095004786]
We propose a dual-agent safe reinforcement learning strategy consisting of a baseline and a safe agent.
Such a decoupled framework enables high flexibility, data efficiency and risk-awareness for RL-based control.
The proposed method outperforms the state-of-the-art safe RL algorithms on difficult robot locomotion and manipulation tasks.
arXiv Detail & Related papers (2022-12-14T03:11:25Z) - Evaluating Model-free Reinforcement Learning toward Safety-critical
Tasks [70.76757529955577]
This paper revisits prior work in this scope from the perspective of state-wise safe RL.
We propose Unrolling Safety Layer (USL), a joint method that combines safety optimization and safety projection.
To facilitate further research in this area, we reproduce related algorithms in a unified pipeline and incorporate them into SafeRL-Kit.
arXiv Detail & Related papers (2022-12-12T06:30:17Z) - Recursively Feasible Probabilistic Safe Online Learning with Control Barrier Functions [60.26921219698514]
We introduce a model-uncertainty-aware reformulation of CBF-based safety-critical controllers.
We then present the pointwise feasibility conditions of the resulting safety controller.
We use these conditions to devise an event-triggered online data collection strategy.
arXiv Detail & Related papers (2022-08-23T05:02:09Z) - Actor-Critic based Improper Reinforcement Learning [61.430513757337486]
We consider an improper reinforcement learning setting where a learner is given $M$ base controllers for an unknown Markov decision process.
We propose two algorithms: (1) a Policy Gradient-based approach; and (2) an algorithm that can switch between a simple Actor-Critic scheme and a Natural Actor-Critic scheme.
arXiv Detail & Related papers (2022-07-19T05:55:02Z) - Safe Pontryagin Differentiable Programming [17.63374326658473]
We propose a theoretical and algorithmic safe differentiable framework to solve a broad class of safety-critical learning and control tasks.
We demonstrate the capabilities of Safe PDP in solving various safe learning and control tasks.
arXiv Detail & Related papers (2021-05-31T13:03:00Z) - Towards Safe Continuing Task Reinforcement Learning [21.390201009230246]
We propose an algorithm capable of operating in the continuing task setting without the need of restarts.
We evaluate our approach in a numerical example, which shows the capabilities of the proposed approach in learning safe policies via safe exploration.
arXiv Detail & Related papers (2021-02-24T22:12:25Z) - Closing the Closed-Loop Distribution Shift in Safe Imitation Learning [80.05727171757454]
We treat safe optimization-based control strategies as experts in an imitation learning problem.
We train a learned policy that can be cheaply evaluated at run-time and that provably satisfies the same safety guarantees as the expert.
arXiv Detail & Related papers (2021-02-18T05:11:41Z) - Runtime-Safety-Guided Policy Repair [13.038017178545728]
We study the problem of policy repair for learning-based control policies in safety-critical settings.
We propose to reduce or even eliminate control switching by repairing' the trained policy based on runtime data produced by the safety controller.
arXiv Detail & Related papers (2020-08-17T23:31:48Z) - Safe Reinforcement Learning via Curriculum Induction [94.67835258431202]
In safety-critical applications, autonomous agents may need to learn in an environment where mistakes can be very costly.
Existing safe reinforcement learning methods make an agent rely on priors that let it avoid dangerous situations.
This paper presents an alternative approach inspired by human teaching, where an agent learns under the supervision of an automatic instructor.
arXiv Detail & Related papers (2020-06-22T10:48:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.