Learning Accurate and Human-Like Driving using Semantic Maps and
Attention
- URL: http://arxiv.org/abs/2007.07218v1
- Date: Fri, 10 Jul 2020 22:25:27 GMT
- Title: Learning Accurate and Human-Like Driving using Semantic Maps and
Attention
- Authors: Simon Hecker, Dengxin Dai, Alexander Liniger, Luc Van Gool
- Abstract summary: This paper investigates how end-to-end driving models can be improved to drive more accurately and human-like.
We exploit semantic and visual maps from HERE Technologies and augment the existing Drive360 dataset with such.
Our models are trained and evaluated on the Drive360 + HERE dataset, which features 60 hours and 3000 km of real-world driving data.
- Score: 152.48143666881418
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper investigates how end-to-end driving models can be improved to
drive more accurately and human-like. To tackle the first issue we exploit
semantic and visual maps from HERE Technologies and augment the existing
Drive360 dataset with such. The maps are used in an attention mechanism that
promotes segmentation confidence masks, thus focusing the network on semantic
classes in the image that are important for the current driving situation.
Human-like driving is achieved using adversarial learning, by not only
minimizing the imitation loss with respect to the human driver but by further
defining a discriminator, that forces the driving model to produce action
sequences that are human-like. Our models are trained and evaluated on the
Drive360 + HERE dataset, which features 60 hours and 3000 km of real-world
driving data. Extensive experiments show that our driving models are more
accurate and behave more human-like than previous methods.
Related papers
- Guiding Attention in End-to-End Driving Models [49.762868784033785]
Vision-based end-to-end driving models trained by imitation learning can lead to affordable solutions for autonomous driving.
We study how to guide the attention of these models to improve their driving quality by adding a loss term during training.
In contrast to previous work, our method does not require these salient semantic maps to be available during testing time.
arXiv Detail & Related papers (2024-04-30T23:18:51Z) - Scaling Vision-based End-to-End Driving with Multi-View Attention
Learning [7.14967754486195]
We present CIL++, which improves on CILRS by both processing higher-resolution images using a human-inspired HFOV as an inductive bias and incorporating a proper attention mechanism.
We propose to replace CILRS with CIL++ as a strong vision-based pure end-to-end driving baseline supervised by only vehicle signals and trained by conditional imitation learning.
arXiv Detail & Related papers (2023-02-07T02:14:45Z) - Policy Pre-training for End-to-end Autonomous Driving via
Self-supervised Geometric Modeling [96.31941517446859]
We propose PPGeo (Policy Pre-training via Geometric modeling), an intuitive and straightforward fully self-supervised framework curated for the policy pretraining in visuomotor driving.
We aim at learning policy representations as a powerful abstraction by modeling 3D geometric scenes on large-scale unlabeled and uncalibrated YouTube driving videos.
In the first stage, the geometric modeling framework generates pose and depth predictions simultaneously, with two consecutive frames as input.
In the second stage, the visual encoder learns driving policy representation by predicting the future ego-motion and optimizing with the photometric error based on current visual observation only.
arXiv Detail & Related papers (2023-01-03T08:52:49Z) - FBLNet: FeedBack Loop Network for Driver Attention Prediction [75.83518507463226]
Nonobjective driving experience is difficult to model.
In this paper, we propose a FeedBack Loop Network (FBLNet) which attempts to model the driving experience accumulation procedure.
Under the guidance of the incremental knowledge, our model fuses the CNN feature and Transformer feature that are extracted from the input image to predict driver attention.
arXiv Detail & Related papers (2022-12-05T08:25:09Z) - Exploring the trade off between human driving imitation and safety for
traffic simulation [0.34410212782758043]
We show that a trade-off exists between imitating human driving and maintaining safety when learning driving policies.
We propose a multi objective learning algorithm (MOPPO) that improves both objectives together.
arXiv Detail & Related papers (2022-08-09T14:30:19Z) - Online Adaptation of Parameters using GRU-based Neural Network with BO
for Accurate Driving Model [0.8433000039153409]
Calibrating a driving model (DM) makes the simulated driving behavior closer to human-driving behavior.
Conventional DM-calibrating methods do not take into account that the parameters in a DM vary while driving.
We propose a DM-calibration method for measuring human driving styles to reproduce real car-following behavior more accurately.
arXiv Detail & Related papers (2021-09-24T03:07:12Z) - IntentNet: Learning to Predict Intention from Raw Sensor Data [86.74403297781039]
In this paper, we develop a one-stage detector and forecaster that exploits both 3D point clouds produced by a LiDAR sensor as well as dynamic maps of the environment.
Our multi-task model achieves better accuracy than the respective separate modules while saving computation, which is critical to reducing reaction time in self-driving applications.
arXiv Detail & Related papers (2021-01-20T00:31:52Z) - Fine-Grained Vehicle Perception via 3D Part-Guided Visual Data
Augmentation [77.60050239225086]
We propose an effective training data generation process by fitting a 3D car model with dynamic parts to vehicles in real images.
Our approach is fully automatic without any human interaction.
We present a multi-task network for VUS parsing and a multi-stream network for VHI parsing.
arXiv Detail & Related papers (2020-12-15T03:03:38Z) - Action-Based Representation Learning for Autonomous Driving [8.296684637620551]
We propose to use action-based driving data for learning representations.
Our experiments show that an affordance-based driving model pre-trained with this approach can leverage a relatively small amount of weakly annotated imagery.
arXiv Detail & Related papers (2020-08-21T10:49:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.