PreTraM: Self-Supervised Pre-training via Connecting Trajectory and Map
- URL: http://arxiv.org/abs/2204.10435v1
- Date: Thu, 21 Apr 2022 23:01:21 GMT
- Title: PreTraM: Self-Supervised Pre-training via Connecting Trajectory and Map
- Authors: Chenfeng Xu, Tian Li, Chen Tang, Lingfeng Sun, Kurt Keutzer, Masayoshi
Tomizuka, Alireza Fathi, Wei Zhan
- Abstract summary: We propose PreTraM, a self-supervised pre-training scheme for trajectory forecasting.
It consists of two parts: 1) Trajectory-Map Contrastive Learning, where we project trajectories and maps to a shared embedding space with cross-modal contrastive learning, and 2) Map Contrastive Learning, where we enhance map representation with contrastive learning on large quantities of HD-maps.
On top of popular baselines such as AgentFormer and Trajectron++, PreTraM boosts their performance by 5.5% and 6.9% relatively in FDE-10 on the challenging nuScenes dataset.
- Score: 58.53373202647576
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep learning has recently achieved significant progress in trajectory
forecasting. However, the scarcity of trajectory data inhibits the data-hungry
deep-learning models from learning good representations. While mature
representation learning methods exist in computer vision and natural language
processing, these pre-training methods require large-scale data. It is hard to
replicate these approaches in trajectory forecasting due to the lack of
adequate trajectory data (e.g., 34K samples in the nuScenes dataset). To work
around the scarcity of trajectory data, we resort to another data modality
closely related to trajectories-HD-maps, which is abundantly provided in
existing datasets. In this paper, we propose PreTraM, a self-supervised
pre-training scheme via connecting trajectories and maps for trajectory
forecasting. Specifically, PreTraM consists of two parts: 1) Trajectory-Map
Contrastive Learning, where we project trajectories and maps to a shared
embedding space with cross-modal contrastive learning, and 2) Map Contrastive
Learning, where we enhance map representation with contrastive learning on
large quantities of HD-maps. On top of popular baselines such as AgentFormer
and Trajectron++, PreTraM boosts their performance by 5.5% and 6.9% relatively
in FDE-10 on the challenging nuScenes dataset. We show that PreTraM improves
data efficiency and scales well with model size.
Related papers
- T-JEPA: A Joint-Embedding Predictive Architecture for Trajectory Similarity Computation [6.844357745770191]
Trajectory similarity computation is an essential technique for analyzing moving patterns of spatial data across various applications.
We propose T-JEPA, a self-supervised trajectory similarity method employing Joint-Embedding Predictive Architecture (JEPA) to enhance trajectory representation learning.
arXiv Detail & Related papers (2024-06-13T09:51:51Z) - Rethinking Transformers Pre-training for Multi-Spectral Satellite
Imagery [78.43828998065071]
Recent advances in unsupervised learning have demonstrated the ability of large vision models to achieve promising results on downstream tasks.
Such pre-training techniques have also been explored recently in the remote sensing domain due to the availability of large amount of unlabelled data.
In this paper, we re-visit transformers pre-training and leverage multi-scale information that is effectively utilized with multiple modalities.
arXiv Detail & Related papers (2024-03-08T16:18:04Z) - Towards A Foundation Model For Trajectory Intelligence [0.0]
We present the results of training a large trajectory model using real-world user check-in data.
Our approach follows a pre-train and fine-tune paradigm, where a base model is pre-trained via masked trajectory modeling.
Our empirical analysis utilizes a comprehensive dataset of over 2 billion check-ins generated by more than 6 million users.
arXiv Detail & Related papers (2023-11-30T00:34:09Z) - SPOT: Scalable 3D Pre-training via Occupancy Prediction for Learning Transferable 3D Representations [76.45009891152178]
Pretraining-finetuning approach can alleviate the labeling burden by fine-tuning a pre-trained backbone across various downstream datasets as well as tasks.
We show, for the first time, that general representations learning can be achieved through the task of occupancy prediction.
Our findings will facilitate the understanding of LiDAR points and pave the way for future advancements in LiDAR pre-training.
arXiv Detail & Related papers (2023-09-19T11:13:01Z) - Pre-training on Synthetic Driving Data for Trajectory Prediction [61.520225216107306]
We propose a pipeline-level solution to mitigate the issue of data scarcity in trajectory forecasting.
We adopt HD map augmentation and trajectory synthesis for generating driving data, and then we learn representations by pre-training on them.
We conduct extensive experiments to demonstrate the effectiveness of our data expansion and pre-training strategies.
arXiv Detail & Related papers (2023-09-18T19:49:22Z) - Minimizing the Accumulated Trajectory Error to Improve Dataset
Distillation [151.70234052015948]
We propose a novel approach that encourages the optimization algorithm to seek a flat trajectory.
We show that the weights trained on synthetic data are robust against the accumulated errors perturbations with the regularization towards the flat trajectory.
Our method, called Flat Trajectory Distillation (FTD), is shown to boost the performance of gradient-matching methods by up to 4.7%.
arXiv Detail & Related papers (2022-11-20T15:49:11Z) - Generating Synthetic Training Data for Deep Learning-Based UAV
Trajectory Prediction [11.241614693184323]
We present an approach for generating synthetic trajectory data of unmanned-aerial-vehicles (UAVs) in image space.
We show that an RNN-based prediction model solely trained on the generated data can outperform classic reference models on a real-world UAV tracking dataset.
arXiv Detail & Related papers (2021-07-01T13:08:31Z) - Injecting Knowledge in Data-driven Vehicle Trajectory Predictors [82.91398970736391]
Vehicle trajectory prediction tasks have been commonly tackled from two perspectives: knowledge-driven or data-driven.
In this paper, we propose to learn a "Realistic Residual Block" (RRB) which effectively connects these two perspectives.
Our proposed method outputs realistic predictions by confining the residual range and taking into account its uncertainty.
arXiv Detail & Related papers (2021-03-08T16:03:09Z) - Mobility Inference on Long-Tailed Sparse Trajectory [2.4444287331956898]
We propose a single trajectory inference algorithm that utilizes a generic long-tailed sparsity pattern in the large-scale trajectory data.
The algorithm guarantees a 100% precision in the stay/travel inference with a provable lower-bound in the recall.
Evaluations with three trajectory data sets of 40 million urban users validate the performance guarantees of the proposed inference algorithm.
arXiv Detail & Related papers (2020-01-21T16:32:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.