Verifying Learning-Based Robotic Navigation Systems
- URL: http://arxiv.org/abs/2205.13536v1
- Date: Thu, 26 May 2022 17:56:43 GMT
- Title: Verifying Learning-Based Robotic Navigation Systems
- Authors: Guy Amir, Davide Corsi, Raz Yerushalmi, Luca Marzari, David Harel,
Alessandro Farinelli and Guy Katz
- Abstract summary: We show how modern verification engines can be used for effective model selection.
Specifically, we use verification to detect and rule out policies that may demonstrate suboptimal behavior.
Our work is the first to demonstrate the use of verification backends for recognizing suboptimal DRL policies in real-world robots.
- Score: 61.01217374879221
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep reinforcement learning (DRL) has become a dominant deep-learning
paradigm for various tasks in which complex policies are learned within
reactive systems. In parallel, there has recently been significant research on
verifying deep neural networks. However, to date, there has been little work
demonstrating the use of modern verification tools on real, DRL-controlled
systems.
In this case-study paper, we attempt to begin bridging this gap, and focus on
the important task of mapless robotic navigation -- a classic robotics problem,
in which a robot, usually controlled by a DRL agent, needs to efficiently and
safely navigate through an unknown arena towards a desired target. We
demonstrate how modern verification engines can be used for effective model
selection, i.e., the process of selecting the best available policy for the
robot in question from a pool of candidate policies. Specifically, we use
verification to detect and rule out policies that may demonstrate suboptimal
behavior, such as collisions and infinite loops. We also apply verification to
identify models with overly conservative behavior, thus allowing users to
choose superior policies that are better at finding an optimal, shorter path to
a target.
To validate our work, we conducted extensive experiments on an actual robot,
and confirmed that the suboptimal policies detected by our method were indeed
flawed. We also compared our verification-driven approach to state-of-the-art
gradient attacks, and our results demonstrate that gradient-based methods are
inadequate in this setting.
Our work is the first to demonstrate the use of DNN verification backends for
recognizing suboptimal DRL policies in real-world robots, and for filtering out
unwanted policies. We believe that the methods presented in this work can be
applied to a large range of application domains that incorporate
deep-learning-based agents.
Related papers
- SERL: A Software Suite for Sample-Efficient Robotic Reinforcement
Learning [85.21378553454672]
We develop a library containing a sample efficient off-policy deep RL method, together with methods for computing rewards and resetting the environment.
We find that our implementation can achieve very efficient learning, acquiring policies for PCB board assembly, cable routing, and object relocation.
These policies achieve perfect or near-perfect success rates, extreme robustness even under perturbations, and exhibit emergent robustness recovery and correction behaviors.
arXiv Detail & Related papers (2024-01-29T10:01:10Z) - NoMaD: Goal Masked Diffusion Policies for Navigation and Exploration [57.15811390835294]
This paper describes how we can train a single unified diffusion policy to handle both goal-directed navigation and goal-agnostic exploration.
We show that this unified policy results in better overall performance when navigating to visually indicated goals in novel environments.
Our experiments, conducted on a real-world mobile robot platform, show effective navigation in unseen environments in comparison with five alternative methods.
arXiv Detail & Related papers (2023-10-11T21:07:14Z) - Learning to View: Decision Transformers for Active Object Detection [18.211691238072245]
In most robotic systems, perception is typically independent of motion planning.
We use reinforcement learning (RL) methods to control the robot in order to obtain images that maximize the detection quality.
We evaluate the performance of proposed method on an interactive dataset collected from an indoor scenario simulator.
arXiv Detail & Related papers (2023-01-23T17:00:48Z) - Constrained Reinforcement Learning for Robotics via Scenario-Based
Programming [64.07167316957533]
It is crucial to optimize the performance of DRL-based agents while providing guarantees about their behavior.
This paper presents a novel technique for incorporating domain-expert knowledge into a constrained DRL training loop.
Our experiments demonstrate that using our approach to leverage expert knowledge dramatically improves the safety and the performance of the agent.
arXiv Detail & Related papers (2022-06-20T07:19:38Z) - Robot Learning of Mobile Manipulation with Reachability Behavior Priors [38.49783454634775]
Mobile Manipulation (MM) systems are ideal candidates for taking up the role of a personal assistant in unstructured real-world environments.
Among other challenges, MM requires effective coordination of the robot's embodiments for executing tasks that require both mobility and manipulation.
We study the integration of robotic reachability priors in actor-critic RL methods for accelerating the learning of MM for reaching and fetching tasks.
arXiv Detail & Related papers (2022-03-08T12:44:42Z) - A Workflow for Offline Model-Free Robotic Reinforcement Learning [117.07743713715291]
offline reinforcement learning (RL) enables learning control policies by utilizing only prior experience, without any online interaction.
We develop a practical workflow for using offline RL analogous to the relatively well-understood for supervised learning problems.
We demonstrate the efficacy of this workflow in producing effective policies without any online tuning.
arXiv Detail & Related papers (2021-09-22T16:03:29Z) - Direct Random Search for Fine Tuning of Deep Reinforcement Learning
Policies [5.543220407902113]
We show that a direct random search is very effective at fine-tuning DRL policies by directly optimizing them using deterministic rollouts.
Our results show that this method yields more consistent and higher performing agents on the environments we tested.
arXiv Detail & Related papers (2021-09-12T20:12:46Z) - Learning Off-Policy with Online Planning [18.63424441772675]
We investigate a novel instantiation of H-step lookahead with a learned model and a terminal value function.
We show the flexibility of LOOP to incorporate safety constraints during deployment with a set of navigation environments.
arXiv Detail & Related papers (2020-08-23T16:18:44Z) - Guided Uncertainty-Aware Policy Optimization: Combining Learning and
Model-Based Strategies for Sample-Efficient Policy Learning [75.56839075060819]
Traditional robotic approaches rely on an accurate model of the environment, a detailed description of how to perform the task, and a robust perception system to keep track of the current state.
reinforcement learning approaches can operate directly from raw sensory inputs with only a reward signal to describe the task, but are extremely sample-inefficient and brittle.
In this work, we combine the strengths of model-based methods with the flexibility of learning-based methods to obtain a general method that is able to overcome inaccuracies in the robotics perception/actuation pipeline.
arXiv Detail & Related papers (2020-05-21T19:47:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.