PreTR: Spatio-Temporal Non-Autoregressive Trajectory Prediction
Transformer
- URL: http://arxiv.org/abs/2203.09293v1
- Date: Thu, 17 Mar 2022 12:52:23 GMT
- Title: PreTR: Spatio-Temporal Non-Autoregressive Trajectory Prediction
Transformer
- Authors: Lina Achaji, Thierno Barry, Thibault Fouqueray, Julien Moreau,
Francois Aioun, Francois Charpillet
- Abstract summary: We introduce a model called PRediction Transformer (PReTR) that extracts features from the multi-agent scenes by employing a factorized-temporal attention module.
It shows less computational needs than previously studied models with empirically better results.
We leverage encoder-decoder Transformer networks for parallel decoding a set of learned object queries.
- Score: 0.9786690381850356
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Nowadays, our mobility systems are evolving into the era of intelligent
vehicles that aim to improve road safety. Due to their vulnerability,
pedestrians are the users who will benefit the most from these developments.
However, predicting their trajectory is one of the most challenging concerns.
Indeed, accurate prediction requires a good understanding of multi-agent
interactions that can be complex. Learning the underlying spatial and temporal
patterns caused by these interactions is even more of a competitive and open
problem that many researchers are tackling. In this paper, we introduce a model
called PRediction Transformer (PReTR) that extracts features from the
multi-agent scenes by employing a factorized spatio-temporal attention module.
It shows less computational needs than previously studied models with
empirically better results. Besides, previous works in motion prediction suffer
from the exposure bias problem caused by generating future sequences
conditioned on model prediction samples rather than ground-truth samples. In
order to go beyond the proposed solutions, we leverage encoder-decoder
Transformer networks for parallel decoding a set of learned object queries.
This non-autoregressive solution avoids the need for iterative conditioning and
arguably decreases training and testing computational time. We evaluate our
model on the ETH/UCY datasets, a publicly available benchmark for pedestrian
trajectory prediction. Finally, we justify our usage of the parallel decoding
technique by showing that the trajectory prediction task can be better solved
as a non-autoregressive task.
Related papers
- OPUS: Occupancy Prediction Using a Sparse Set [64.60854562502523]
We present a framework to simultaneously predict occupied locations and classes using a set of learnable queries.
OPUS incorporates a suite of non-trivial strategies to enhance model performance.
Our lightest model achieves superior RayIoU on the Occ3D-nuScenes dataset at near 2x FPS, while our heaviest model surpasses previous best results by 6.1 RayIoU.
arXiv Detail & Related papers (2024-09-14T07:44:22Z) - T4P: Test-Time Training of Trajectory Prediction via Masked Autoencoder and Actor-specific Token Memory [39.021321011792786]
Trajectory prediction is a challenging problem that requires considering interactions among multiple actors.
Data-driven approaches have been used to address this complex problem, but they suffer from unreliable predictions under distribution shifts during test time.
We propose several online learning methods using regression loss from the ground truth of observed data.
Our method surpasses the performance of existing state-of-the-art online learning methods in terms of both prediction accuracy and computational efficiency.
arXiv Detail & Related papers (2024-03-15T06:47:14Z) - Attention-aware Social Graph Transformer Networks for Stochastic Trajectory Prediction [16.55909815712467]
Trajectory prediction is fundamental to various intelligent technologies, such as autonomous driving and robotics.
Current trajectory prediction research faces problems of complex social interactions, high dynamics and multi-modality.
We propose Attention-aware Social Graph Transformer Networks for multi-modal trajectory prediction.
arXiv Detail & Related papers (2023-12-26T04:24:01Z) - EANet: Expert Attention Network for Online Trajectory Prediction [5.600280639034753]
Expert Attention Network is a complete online learning framework for trajectory prediction.
We introduce expert attention, which adjusts the weights of different depths of network layers, avoiding the model updated slowly due to gradient problem.
Furthermore, we propose a short-term motion trend kernel function which is sensitive to scenario change, allowing the model to respond quickly.
arXiv Detail & Related papers (2023-09-11T07:09:40Z) - Implicit Occupancy Flow Fields for Perception and Prediction in
Self-Driving [68.95178518732965]
A self-driving vehicle (SDV) must be able to perceive its surroundings and predict the future behavior of other traffic participants.
Existing works either perform object detection followed by trajectory of the detected objects, or predict dense occupancy and flow grids for the whole scene.
This motivates our unified approach to perception and future prediction that implicitly represents occupancy and flow over time with a single neural network.
arXiv Detail & Related papers (2023-08-02T23:39:24Z) - Pedestrian Trajectory Prediction via Spatial Interaction Transformer
Network [7.150832716115448]
In traffic scenes, when encountering with oncoming people, pedestrians may make sudden turns or stop immediately.
To predict such unpredictable trajectories, we can gain insights into the interaction between pedestrians.
We present a novel generative method named Spatial Interaction Transformer (SIT), which learns the correlation of pedestrian trajectories through attention mechanisms.
arXiv Detail & Related papers (2021-12-13T13:08:04Z) - You Mostly Walk Alone: Analyzing Feature Attribution in Trajectory
Prediction [52.442129609979794]
Recent deep learning approaches for trajectory prediction show promising performance.
It remains unclear which features such black-box models actually learn to use for making predictions.
This paper proposes a procedure that quantifies the contributions of different cues to model performance.
arXiv Detail & Related papers (2021-10-11T14:24:15Z) - Learning Sparse Interaction Graphs of Partially Observed Pedestrians for
Trajectory Prediction [0.3025231207150811]
Multi-pedestrian trajectory prediction is an indispensable safety element of autonomous systems that interact with crowds in unstructured environments.
We propose Gumbel Social Transformer, in which an Edge Gumbel Selector samples a sparse graph of partially observed pedestrians at each time step.
We demonstrate that our model overcomes the potential problems caused by the assumptions, and our approach outperforms the related works in benchmark evaluation.
arXiv Detail & Related papers (2021-07-15T00:45:11Z) - SGCN:Sparse Graph Convolution Network for Pedestrian Trajectory
Prediction [64.16212996247943]
We present a Sparse Graph Convolution Network(SGCN) for pedestrian trajectory prediction.
Specifically, the SGCN explicitly models the sparse directed interaction with a sparse directed spatial graph to capture adaptive interaction pedestrians.
visualizations indicate that our method can capture adaptive interactions between pedestrians and their effective motion tendencies.
arXiv Detail & Related papers (2021-04-04T03:17:42Z) - Injecting Knowledge in Data-driven Vehicle Trajectory Predictors [82.91398970736391]
Vehicle trajectory prediction tasks have been commonly tackled from two perspectives: knowledge-driven or data-driven.
In this paper, we propose to learn a "Realistic Residual Block" (RRB) which effectively connects these two perspectives.
Our proposed method outputs realistic predictions by confining the residual range and taking into account its uncertainty.
arXiv Detail & Related papers (2021-03-08T16:03:09Z) - End-to-end Contextual Perception and Prediction with Interaction
Transformer [79.14001602890417]
We tackle the problem of detecting objects in 3D and forecasting their future motion in the context of self-driving.
To capture their spatial-temporal dependencies, we propose a recurrent neural network with a novel Transformer architecture.
Our model can be trained end-to-end, and runs in real-time.
arXiv Detail & Related papers (2020-08-13T14:30:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.