Driving Through Ghosts: Behavioral Cloning with False Positives
- URL: http://arxiv.org/abs/2008.12969v1
- Date: Sat, 29 Aug 2020 12:10:23 GMT
- Title: Driving Through Ghosts: Behavioral Cloning with False Positives
- Authors: Andreas B\"uhler, Adrien Gaidon, Andrei Cramariuc, Rares Ambrus, Guy
Rosman, Wolfram Burgard
- Abstract summary: We propose a behavioral cloning approach that can safely leverage imperfect perception without being conservative.
We propose a new probabilistic birds-eye-view semantic grid to encode the noisy output of object perception systems.
We then leverage expert demonstrations to learn an imitative driving policy using this probabilistic representation.
- Score: 42.31740099795908
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Safe autonomous driving requires robust detection of other traffic
participants. However, robust does not mean perfect, and safe systems typically
minimize missed detections at the expense of a higher false positive rate. This
results in conservative and yet potentially dangerous behavior such as avoiding
imaginary obstacles. In the context of behavioral cloning, perceptual errors at
training time can lead to learning difficulties or wrong policies, as expert
demonstrations might be inconsistent with the perceived world state. In this
work, we propose a behavioral cloning approach that can safely leverage
imperfect perception without being conservative. Our core contribution is a
novel representation of perceptual uncertainty for learning to plan. We propose
a new probabilistic birds-eye-view semantic grid to encode the noisy output of
object perception systems. We then leverage expert demonstrations to learn an
imitative driving policy using this probabilistic representation. Using the
CARLA simulator, we show that our approach can safely overcome critical false
positives that would otherwise lead to catastrophic failures or conservative
behavior.
Related papers
- RACER: Epistemic Risk-Sensitive RL Enables Fast Driving with Fewer Crashes [57.319845580050924]
We propose a reinforcement learning framework that combines risk-sensitive control with an adaptive action space curriculum.
We show that our algorithm is capable of learning high-speed policies for a real-world off-road driving task.
arXiv Detail & Related papers (2024-05-07T23:32:36Z) - SAFE-GIL: SAFEty Guided Imitation Learning [7.979892202477701]
Behavior cloning is a popular approach to Imitation Learning, in which a robot observes an expert supervisor and learns a control policy.
However, behavior cloning suffers from the "compounding error" problem - the policy errors compound as it deviates from the expert demonstrations and might lead to catastrophic system failures.
We propose SAFE-GIL, an off-policy behavior cloning method that guides the expert via adversarial disturbance during data collection.
arXiv Detail & Related papers (2024-04-08T07:25:25Z) - Safety Margins for Reinforcement Learning [53.10194953873209]
We show how to leverage proxy criticality metrics to generate safety margins.
We evaluate our approach on learned policies from APE-X and A3C within an Atari environment.
arXiv Detail & Related papers (2023-07-25T16:49:54Z) - How to Learn from Risk: Explicit Risk-Utility Reinforcement Learning for
Efficient and Safe Driving Strategies [1.496194593196997]
This paper proposes SafeDQN, which allows to make the behavior of autonomous vehicles safe and interpretable while still being efficient.
We show that SafeDQN finds interpretable and safe driving policies for a variety of scenarios and demonstrate how state-of-the-art saliency techniques can help to assess both risk and utility.
arXiv Detail & Related papers (2022-03-16T05:51:22Z) - ProBF: Learning Probabilistic Safety Certificates with Barrier Functions [31.203344483485843]
The control barrier function is a useful tool to guarantee safety if we have access to the ground-truth system dynamics.
In practice, we have inaccurate knowledge of the system dynamics, which can lead to unsafe behaviors.
We show the efficacy of this method through experiments on Segway and Quadrotor simulations.
arXiv Detail & Related papers (2021-12-22T20:18:18Z) - Probabilistic Approach for Road-Users Detection [0.0]
One of the major challenges in state-of-the-art deep-learning based object detection are false positives which occur with overconfident scores.
This paper proposes an approach to alleviate the problem of overconfident predictions by introducing a novel probabilistic layer to deep object detection networks in testing.
It is demonstrated that the proposed technique reduces overconfidence in the false positives without degrading the performance on the true positives.
arXiv Detail & Related papers (2021-12-02T16:02:08Z) - Learning to Be Cautious [71.9871661858886]
A key challenge in the field of reinforcement learning is to develop agents that behave cautiously in novel situations.
We present a sequence of tasks where cautious behavior becomes increasingly non-obvious, as well as an algorithm to demonstrate that it is possible for a system to emphlearn to be cautious.
arXiv Detail & Related papers (2021-10-29T16:52:45Z) - Sample-Efficient Safety Assurances using Conformal Prediction [57.92013073974406]
Early warning systems can provide alerts when an unsafe situation is imminent.
To reliably improve safety, these warning systems should have a provable false negative rate.
We present a framework that combines a statistical inference technique known as conformal prediction with a simulator of robot/environment dynamics.
arXiv Detail & Related papers (2021-09-28T23:00:30Z) - Learning Uncertainty For Safety-Oriented Semantic Segmentation In
Autonomous Driving [77.39239190539871]
We show how uncertainty estimation can be leveraged to enable safety critical image segmentation in autonomous driving.
We introduce a new uncertainty measure based on disagreeing predictions as measured by a dissimilarity function.
We show experimentally that our proposed approach is much less computationally intensive at inference time than competing methods.
arXiv Detail & Related papers (2021-05-28T09:23:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.