Vehicle Trajectory Prediction on Highways Using Bird Eye View
Representations and Deep Learning
- URL: http://arxiv.org/abs/2207.01407v1
- Date: Mon, 4 Jul 2022 13:39:46 GMT
- Title: Vehicle Trajectory Prediction on Highways Using Bird Eye View
Representations and Deep Learning
- Authors: Rub\'en Izquierdo, \'Alvaro Quintanar, David Fern\'andez Llorca,
Iv\'an Garc\'ia Daza, Noelia Hern\'andez, Ignacio Parra, Miguel \'Angel
Sotelo
- Abstract summary: This work presents a novel method for predicting vehicle trajectories in highway scenarios using efficient bird's eye view representations and convolutional neural networks.
The U-net model has been selected as the prediction kernel to generate future visual representations of the scene using an image-to-image regression approach.
A method has been implemented to extract vehicle positions from the generated graphical representations to achieve subpixel resolution.
- Score: 0.5420492913071214
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: This work presents a novel method for predicting vehicle trajectories in
highway scenarios using efficient bird's eye view representations and
convolutional neural networks. Vehicle positions, motion histories, road
configuration, and vehicle interactions are easily included in the prediction
model using basic visual representations. The U-net model has been selected as
the prediction kernel to generate future visual representations of the scene
using an image-to-image regression approach. A method has been implemented to
extract vehicle positions from the generated graphical representations to
achieve subpixel resolution. The method has been trained and evaluated using
the PREVENTION dataset, an on-board sensor dataset. Different network
configurations and scene representations have been evaluated. This study found
that U-net with 6 depth levels using a linear terminal layer and a Gaussian
representation of the vehicles is the best performing configuration. The use of
lane markings was found to produce no improvement in prediction performance.
The average prediction error is 0.47 and 0.38 meters and the final prediction
error is 0.76 and 0.53 meters for longitudinal and lateral coordinates,
respectively, for a predicted trajectory length of 2.0 seconds. The prediction
error is up to 50% lower compared to the baseline method.
Related papers
- VisionTrap: Vision-Augmented Trajectory Prediction Guided by Textual Descriptions [10.748597086208145]
In this work, we propose a novel method that also incorporates visual input from surround-view cameras.
Our method achieves a latency of 53 ms, making it feasible for real-time processing.
Our experiments show that both the visual inputs and the textual descriptions contribute to improvements in trajectory prediction performance.
arXiv Detail & Related papers (2024-07-17T06:39:52Z) - BEVSeg2TP: Surround View Camera Bird's-Eye-View Based Joint Vehicle
Segmentation and Ego Vehicle Trajectory Prediction [4.328789276903559]
Trajectory prediction is a key task for vehicle autonomy.
There is a growing interest in learning-based trajectory prediction.
We show that there is the potential to improve the performance of perception.
arXiv Detail & Related papers (2023-12-20T15:02:37Z) - A Novel Deep Neural Network for Trajectory Prediction in Automated
Vehicles Using Velocity Vector Field [12.067838086415833]
This paper proposes a novel technique for trajectory prediction that combines a data-driven learning-based method with a velocity vector field (VVF) generated from a nature-inspired concept.
The accuracy remains consistent with decreasing observation windows which alleviates the requirement of a long history of past observations for accurate trajectory prediction.
arXiv Detail & Related papers (2023-09-19T22:14:52Z) - Pre-training on Synthetic Driving Data for Trajectory Prediction [61.520225216107306]
We propose a pipeline-level solution to mitigate the issue of data scarcity in trajectory forecasting.
We adopt HD map augmentation and trajectory synthesis for generating driving data, and then we learn representations by pre-training on them.
We conduct extensive experiments to demonstrate the effectiveness of our data expansion and pre-training strategies.
arXiv Detail & Related papers (2023-09-18T19:49:22Z) - SAPI: Surroundings-Aware Vehicle Trajectory Prediction at Intersections [4.982485708779067]
SAPI is a deep learning model to predict vehicle trajectories at intersections.
The proposed model consists of two convolutional network (CNN) and recurrent neural network (RNN)-based encoders and one decoder.
We evaluate SAPI on a proprietary dataset collected in real-world intersections through autonomous vehicles.
arXiv Detail & Related papers (2023-06-02T07:10:45Z) - Multi-Vehicle Trajectory Prediction at Intersections using State and
Intention Information [50.40632021583213]
Traditional approaches to prediction of future trajectory of road agents rely on knowing information about their past trajectory.
This work instead relies on having knowledge of the current state and intended direction to make predictions for multiple vehicles at intersections.
Message passing of this information between the vehicles provides each one of them a more holistic overview of the environment.
arXiv Detail & Related papers (2023-01-06T15:13:23Z) - LOPR: Latent Occupancy PRediction using Generative Models [49.15687400958916]
LiDAR generated occupancy grid maps (L-OGMs) offer a robust bird's eye-view scene representation.
We propose a framework that decouples occupancy prediction into: representation learning and prediction within the learned latent space.
arXiv Detail & Related papers (2022-10-03T22:04:00Z) - SLPC: a VRNN-based approach for stochastic lidar prediction and
completion in autonomous driving [63.87272273293804]
We propose a new LiDAR prediction framework that is based on generative models namely Variational Recurrent Neural Networks (VRNNs)
Our algorithm is able to address the limitations of previous video prediction frameworks when dealing with sparse data by spatially inpainting the depth maps in the upcoming frames.
We present a sparse version of VRNNs and an effective self-supervised training method that does not require any labels.
arXiv Detail & Related papers (2021-02-19T11:56:44Z) - Vehicle Trajectory Prediction in Crowded Highway Scenarios Using Bird
Eye View Representations and CNNs [0.0]
This paper describes a novel approach to perform vehicle trajectory predictions employing graphic representations.
The problem is faced as an image to image regression problem training the network to learn the underlying relations between the traffic participants.
The model has been tested in highway scenarios with more than 30 vehicles simultaneously in two opposite traffic flow streams.
arXiv Detail & Related papers (2020-08-26T11:15:49Z) - VectorNet: Encoding HD Maps and Agent Dynamics from Vectorized
Representation [74.56282712099274]
This paper introduces VectorNet, a hierarchical graph neural network that exploits the spatial locality of individual road components represented by vectors.
By operating on the vectorized high definition (HD) maps and agent trajectories, we avoid lossy rendering and computationally intensive ConvNet encoding steps.
We evaluate VectorNet on our in-house behavior prediction benchmark and the recently released Argoverse forecasting dataset.
arXiv Detail & Related papers (2020-05-08T19:07:03Z) - Spatiotemporal Relationship Reasoning for Pedestrian Intent Prediction [57.56466850377598]
Reasoning over visual data is a desirable capability for robotics and vision-based applications.
In this paper, we present a framework on graph to uncover relationships in different objects in the scene for reasoning about pedestrian intent.
Pedestrian intent, defined as the future action of crossing or not-crossing the street, is a very crucial piece of information for autonomous vehicles.
arXiv Detail & Related papers (2020-02-20T18:50:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.