Constrained Reinforcement Learning for Robotics via Scenario-Based
Programming
- URL: http://arxiv.org/abs/2206.09603v1
- Date: Mon, 20 Jun 2022 07:19:38 GMT
- Title: Constrained Reinforcement Learning for Robotics via Scenario-Based
Programming
- Authors: Davide Corsi, Raz Yerushalmi, Guy Amir, Alessandro Farinelli, David
Harel, Guy Katz
- Abstract summary: It is crucial to optimize the performance of DRL-based agents while providing guarantees about their behavior.
This paper presents a novel technique for incorporating domain-expert knowledge into a constrained DRL training loop.
Our experiments demonstrate that using our approach to leverage expert knowledge dramatically improves the safety and the performance of the agent.
- Score: 64.07167316957533
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep reinforcement learning (DRL) has achieved groundbreaking successes in a
wide variety of robotic applications. A natural consequence is the adoption of
this paradigm for safety-critical tasks, where human safety and expensive
hardware can be involved. In this context, it is crucial to optimize the
performance of DRL-based agents while providing guarantees about their
behavior. This paper presents a novel technique for incorporating domain-expert
knowledge into a constrained DRL training loop. Our technique exploits the
scenario-based programming paradigm, which is designed to allow specifying such
knowledge in a simple and intuitive way. We validated our method on the popular
robotic mapless navigation problem, in simulation, and on the actual platform.
Our experiments demonstrate that using our approach to leverage expert
knowledge dramatically improves the safety and the performance of the agent.
Related papers
- Aquatic Navigation: A Challenging Benchmark for Deep Reinforcement Learning [53.3760591018817]
We propose a new benchmarking environment for aquatic navigation using recent advances in the integration between game engines and Deep Reinforcement Learning.
Specifically, we focus on PPO, one of the most widely accepted algorithms, and we propose advanced training techniques.
Our empirical evaluation shows that a well-designed combination of these ingredients can achieve promising results.
arXiv Detail & Related papers (2024-05-30T23:20:23Z) - SERL: A Software Suite for Sample-Efficient Robotic Reinforcement
Learning [85.21378553454672]
We develop a library containing a sample efficient off-policy deep RL method, together with methods for computing rewards and resetting the environment.
We find that our implementation can achieve very efficient learning, acquiring policies for PCB board assembly, cable routing, and object relocation.
These policies achieve perfect or near-perfect success rates, extreme robustness even under perturbations, and exhibit emergent robustness recovery and correction behaviors.
arXiv Detail & Related papers (2024-01-29T10:01:10Z) - Evaluating Model-free Reinforcement Learning toward Safety-critical
Tasks [70.76757529955577]
This paper revisits prior work in this scope from the perspective of state-wise safe RL.
We propose Unrolling Safety Layer (USL), a joint method that combines safety optimization and safety projection.
To facilitate further research in this area, we reproduce related algorithms in a unified pipeline and incorporate them into SafeRL-Kit.
arXiv Detail & Related papers (2022-12-12T06:30:17Z) - Verifying Learning-Based Robotic Navigation Systems [61.01217374879221]
We show how modern verification engines can be used for effective model selection.
Specifically, we use verification to detect and rule out policies that may demonstrate suboptimal behavior.
Our work is the first to demonstrate the use of verification backends for recognizing suboptimal DRL policies in real-world robots.
arXiv Detail & Related papers (2022-05-26T17:56:43Z) - Scenario-Assisted Deep Reinforcement Learning [3.5036351567024275]
We propose a technique for enhancing the reinforcement learning training process.
It allows human engineers to directly contribute their expert knowledge, making the agent under training more likely to comply with relevant constraints.
We evaluate our technique using a case-study from the domain of internet congestion control.
arXiv Detail & Related papers (2022-02-09T08:46:13Z) - Accelerating Robotic Reinforcement Learning via Parameterized Action
Primitives [92.0321404272942]
Reinforcement learning can be used to build general-purpose robotic systems.
However, training RL agents to solve robotics tasks still remains challenging.
In this work, we manually specify a library of robot action primitives (RAPS), parameterized with arguments that are learned by an RL policy.
We find that our simple change to the action interface substantially improves both the learning efficiency and task performance.
arXiv Detail & Related papers (2021-10-28T17:59:30Z) - Human-Aware Robot Navigation via Reinforcement Learning with Hindsight
Experience Replay and Curriculum Learning [28.045441768064215]
Reinforcement learning approaches have shown superior ability in solving sequential decision making problems.
In this work, we consider the task of training an RL agent without employing the demonstration data.
We propose to incorporate the hindsight experience replay (HER) and curriculum learning (CL) techniques with RL to efficiently learn the optimal navigation policy in the dense crowd.
arXiv Detail & Related papers (2021-10-09T13:18:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.