Exploring Map-based Features for Efficient Attention-based Vehicle
Motion Prediction
- URL: http://arxiv.org/abs/2205.13071v1
- Date: Wed, 25 May 2022 22:38:11 GMT
- Title: Exploring Map-based Features for Efficient Attention-based Vehicle
Motion Prediction
- Authors: Carlos G\'omez-Hu\'elamo, Marcos V. Conde, Miguel Ortiz
- Abstract summary: Motion prediction of multiple agents is a crucial task in arbitrarily complex environments.
We show how to achieve competitive performance on the Argoverse 1.0 Benchmark using efficient attention-based models.
- Score: 3.222802562733787
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Motion prediction (MP) of multiple agents is a crucial task in arbitrarily
complex environments, from social robots to self-driving cars. Current
approaches tackle this problem using end-to-end networks, where the input data
is usually a rendered top-view of the scene and the past trajectories of all
the agents; leveraging this information is a must to obtain optimal
performance. In that sense, a reliable Autonomous Driving (AD) system must
produce reasonable predictions on time, however, despite many of these
approaches use simple ConvNets and LSTMs, models might not be efficient enough
for real-time applications when using both sources of information (map and
trajectory history). Moreover, the performance of these models highly depends
on the amount of training data, which can be expensive (particularly the
annotated HD maps). In this work, we explore how to achieve competitive
performance on the Argoverse 1.0 Benchmark using efficient attention-based
models, which take as input the past trajectories and map-based features from
minimal map information to ensure efficient and reliable MP. These features
represent interpretable information as the driveable area and plausible goal
points, in opposition to black-box CNN-based methods for map processing.
Related papers
- MFTraj: Map-Free, Behavior-Driven Trajectory Prediction for Autonomous Driving [15.965681867350215]
This paper introduces a trajectory prediction model tailored for autonomous driving.
It harnesses historical trajectory data combined with a novel geometric dynamic graph-based behavior-aware module.
It achieves computational efficiency and reduced parameter overhead.
arXiv Detail & Related papers (2024-05-02T13:13:52Z) - Guiding Attention in End-to-End Driving Models [49.762868784033785]
Vision-based end-to-end driving models trained by imitation learning can lead to affordable solutions for autonomous driving.
We study how to guide the attention of these models to improve their driving quality by adding a loss term during training.
In contrast to previous work, our method does not require these salient semantic maps to be available during testing time.
arXiv Detail & Related papers (2024-04-30T23:18:51Z) - SCOUT+: Towards Practical Task-Driven Drivers' Gaze Prediction [12.246649738388388]
SCOUT+ is a task- and context-aware model for drivers' gaze prediction.
We evaluate our model on two datasets, DR(eye)VE and BDD-A.
arXiv Detail & Related papers (2024-04-12T18:29:10Z) - Efficient Baselines for Motion Prediction in Autonomous Driving [7.608073471097835]
Motion Prediction (MP) of multiple surroundings agents is a crucial task in arbitrarily complex environments.
We aim to develop compact models using State-Of-The-Art (SOTA) techniques for MP, including attention mechanisms and GNNs.
arXiv Detail & Related papers (2023-09-06T22:18:16Z) - Implicit Occupancy Flow Fields for Perception and Prediction in
Self-Driving [68.95178518732965]
A self-driving vehicle (SDV) must be able to perceive its surroundings and predict the future behavior of other traffic participants.
Existing works either perform object detection followed by trajectory of the detected objects, or predict dense occupancy and flow grids for the whole scene.
This motivates our unified approach to perception and future prediction that implicitly represents occupancy and flow over time with a single neural network.
arXiv Detail & Related papers (2023-08-02T23:39:24Z) - A Fast and Map-Free Model for Trajectory Prediction in Traffics [2.435517936694533]
This paper proposes an efficient trajectory prediction model that is not dependent on traffic maps.
By comprehensively utilizing attention mechanism, LSTM, graph convolution network and temporal transformer, our model is able to learn rich dynamic and interaction information of all agents.
Our model achieves the highest performance when comparing with existing map-free methods and also exceeds most map-based state-of-the-art methods on the Argoverse dataset.
arXiv Detail & Related papers (2023-07-19T08:36:31Z) - Video Killed the HD-Map: Predicting Multi-Agent Behavior Directly From
Aerial Images [14.689298253430568]
We propose an aerial image-based map (AIM) representation that requires minimal annotation and provides rich road context information for traffic agents like pedestrians and vehicles.
Our results demonstrate competitive multi-agent trajectory prediction performance especially for pedestrians in the scene when using our AIM representation.
arXiv Detail & Related papers (2023-05-19T17:48:01Z) - PreTraM: Self-Supervised Pre-training via Connecting Trajectory and Map [58.53373202647576]
We propose PreTraM, a self-supervised pre-training scheme for trajectory forecasting.
It consists of two parts: 1) Trajectory-Map Contrastive Learning, where we project trajectories and maps to a shared embedding space with cross-modal contrastive learning, and 2) Map Contrastive Learning, where we enhance map representation with contrastive learning on large quantities of HD-maps.
On top of popular baselines such as AgentFormer and Trajectron++, PreTraM boosts their performance by 5.5% and 6.9% relatively in FDE-10 on the challenging nuScenes dataset.
arXiv Detail & Related papers (2022-04-21T23:01:21Z) - Learnable Online Graph Representations for 3D Multi-Object Tracking [156.58876381318402]
We propose a unified and learning based approach to the 3D MOT problem.
We employ a Neural Message Passing network for data association that is fully trainable.
We show the merit of the proposed approach on the publicly available nuScenes dataset by achieving state-of-the-art performance of 65.6% AMOTA and 58% fewer ID-switches.
arXiv Detail & Related papers (2021-04-23T17:59:28Z) - DS-Net: Dynamic Spatiotemporal Network for Video Salient Object
Detection [78.04869214450963]
We propose a novel dynamic temporal-temporal network (DSNet) for more effective fusion of temporal and spatial information.
We show that the proposed method achieves superior performance than state-of-the-art algorithms.
arXiv Detail & Related papers (2020-12-09T06:42:30Z) - PnPNet: End-to-End Perception and Prediction with Tracking in the Loop [82.97006521937101]
We tackle the problem of joint perception and motion forecasting in the context of self-driving vehicles.
We propose Net, an end-to-end model that takes as input sensor data, and outputs at each time step object tracks and their future level.
arXiv Detail & Related papers (2020-05-29T17:57:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.