Conditional Unscented Autoencoders for Trajectory Prediction
- URL: http://arxiv.org/abs/2310.19944v2
- Date: Tue, 27 Feb 2024 08:16:32 GMT
- Title: Conditional Unscented Autoencoders for Trajectory Prediction
- Authors: Faris Janjo\v{s}, Marcel Hallgarten, Anthony Knittel, Maxim Dolgov,
Andreas Zell, J. Marius Z\"ollner
- Abstract summary: The CVAE is one of the most widely-used models in trajectory prediction for AD.
We leverage recent advances in the space of the VAE, the foundation of the CVAE, which show that a simple change in the sampling procedure can greatly benefit performance.
We show wide applicability of our models by evaluating them on the INTERACTION prediction dataset, outperforming the state of the art, as well as at the task of image modeling on the CelebA dataset.
- Score: 13.121738145903532
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The CVAE is one of the most widely-used models in trajectory prediction for
AD. It captures the interplay between a driving context and its ground-truth
future into a probabilistic latent space and uses it to produce predictions. In
this paper, we challenge key components of the CVAE. We leverage recent
advances in the space of the VAE, the foundation of the CVAE, which show that a
simple change in the sampling procedure can greatly benefit performance. We
find that unscented sampling, which draws samples from any learned distribution
in a deterministic manner, can naturally be better suited to trajectory
prediction than potentially dangerous random sampling. We go further and offer
additional improvements including a more structured Gaussian mixture latent
space, as well as a novel, potentially more expressive way to do inference with
CVAEs. We show wide applicability of our models by evaluating them on the
INTERACTION prediction dataset, outperforming the state of the art, as well as
at the task of image modeling on the CelebA dataset, outperforming the baseline
vanilla CVAE. Code is available at
https://github.com/boschresearch/cuae-prediction.
Related papers
- OPUS: Occupancy Prediction Using a Sparse Set [64.60854562502523]
We present a framework to simultaneously predict occupied locations and classes using a set of learnable queries.
OPUS incorporates a suite of non-trivial strategies to enhance model performance.
Our lightest model achieves superior RayIoU on the Occ3D-nuScenes dataset at near 2x FPS, while our heaviest model surpasses previous best results by 6.1 RayIoU.
arXiv Detail & Related papers (2024-09-14T07:44:22Z) - Learning Augmentation Policies from A Model Zoo for Time Series Forecasting [58.66211334969299]
We introduce AutoTSAug, a learnable data augmentation method based on reinforcement learning.
By augmenting the marginal samples with a learnable policy, AutoTSAug substantially improves forecasting performance.
arXiv Detail & Related papers (2024-09-10T07:34:19Z) - Optimizing Diffusion Models for Joint Trajectory Prediction and Controllable Generation [49.49868273653921]
Diffusion models are promising for joint trajectory prediction and controllable generation in autonomous driving.
We introduce Optimal Gaussian Diffusion (OGD) and Estimated Clean Manifold (ECM) Guidance.
Our methodology streamlines the generative process, enabling practical applications with reduced computational overhead.
arXiv Detail & Related papers (2024-08-01T17:59:59Z) - Improving Out-of-Distribution Generalization of Trajectory Prediction for Autonomous Driving via Polynomial Representations [16.856874154363588]
We present an OoD testing protocol that homogenizes datasets and prediction tasks across two large-scale motion datasets.
With a much smaller model size, training effort, and inference time, we reach near SotA performance for ID testing and significantly improve robustness in OoD testing.
arXiv Detail & Related papers (2024-07-18T12:00:32Z) - Foundation Models for Generalist Geospatial Artificial Intelligence [3.7002058945990415]
This paper introduces a first-of-a-kind framework for the efficient pre-training and fine-tuning of foundational models on extensive data.
We have utilized this framework to create Prithvi, a transformer-based foundational model pre-trained on more than 1TB of multispectral satellite imagery.
arXiv Detail & Related papers (2023-10-28T10:19:55Z) - EquiDiff: A Conditional Equivariant Diffusion Model For Trajectory
Prediction [11.960234424309265]
We propose EquiDiff, a deep generative model for predicting future vehicle trajectories.
EquiDiff is based on the conditional diffusion model, which generates future trajectories by incorporating historical information and random Gaussian noise.
Our results demonstrate that EquiDiff outperforms other baseline models in short-term prediction, but has slightly higher errors for long-term prediction.
arXiv Detail & Related papers (2023-08-12T13:17:09Z) - Prediction-Oriented Bayesian Active Learning [51.426960808684655]
Expected predictive information gain (EPIG) is an acquisition function that measures information gain in the space of predictions rather than parameters.
EPIG leads to stronger predictive performance compared with BALD across a range of datasets and models.
arXiv Detail & Related papers (2023-04-17T10:59:57Z) - Conditioned Human Trajectory Prediction using Iterative Attention Blocks [70.36888514074022]
We present a simple yet effective pedestrian trajectory prediction model aimed at pedestrians positions prediction in urban-like environments.
Our model is a neural-based architecture that can run several layers of attention blocks and transformers in an iterative sequential fashion.
We show that without explicit introduction of social masks, dynamical models, social pooling layers, or complicated graph-like structures, it is possible to produce on par results with SoTA models.
arXiv Detail & Related papers (2022-06-29T07:49:48Z) - Leveraging Unlabeled Data to Predict Out-of-Distribution Performance [63.740181251997306]
Real-world machine learning deployments are characterized by mismatches between the source (training) and target (test) distributions.
In this work, we investigate methods for predicting the target domain accuracy using only labeled source data and unlabeled target data.
We propose Average Thresholded Confidence (ATC), a practical method that learns a threshold on the model's confidence, predicting accuracy as the fraction of unlabeled examples.
arXiv Detail & Related papers (2022-01-11T23:01:12Z) - Adversarial Gradient Driven Exploration for Deep Click-Through Rate
Prediction [39.61776002290324]
We propose a novel exploration method called textbfAdrial textbfGradientversa Driven textbfExploration (AGE)
AGE simulates the gradient updating process, which can approximate the influence of the samples of to-be-explored items for the model.
The effectiveness of our approach was demonstrated on an open-access academic dataset.
arXiv Detail & Related papers (2021-12-21T12:13:07Z) - Exponentially Tilted Gaussian Prior for Variational Autoencoder [3.52359746858894]
Recent studies show that probabilistic generative models can perform poorly on this task.
We propose the exponentially tilted Gaussian prior distribution for the Variational Autoencoder (VAE)
We show that our model produces high quality image samples which are more crisp than that of a standard Gaussian VAE.
arXiv Detail & Related papers (2021-11-30T18:28:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.