EDA: Evolving and Distinct Anchors for Multimodal Motion Prediction
- URL: http://arxiv.org/abs/2312.09501v1
- Date: Fri, 15 Dec 2023 02:55:24 GMT
- Title: EDA: Evolving and Distinct Anchors for Multimodal Motion Prediction
- Authors: Longzhong Lin, Xuewu Lin, Tianwei Lin, Lichao Huang, Rong Xiong, Yue
Wang
- Abstract summary: We introduce a novel paradigm, named Evolving and Distinct Anchors (EDA), to define the positive and negative components for multimodal motion prediction based on mixture models.
EDA enables anchors to evolve and redistribute themselves under specific scenes for an enlarged regression capacity.
- Score: 27.480524917596565
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Motion prediction is a crucial task in autonomous driving, and one of its
major challenges lands in the multimodality of future behaviors. Many
successful works have utilized mixture models which require identification of
positive mixture components, and correspondingly fall into two main lines:
prediction-based and anchor-based matching. The prediction clustering
phenomenon in prediction-based matching makes it difficult to pick
representative trajectories for downstream tasks, while the anchor-based
matching suffers from a limited regression capability. In this paper, we
introduce a novel paradigm, named Evolving and Distinct Anchors (EDA), to
define the positive and negative components for multimodal motion prediction
based on mixture models. We enable anchors to evolve and redistribute
themselves under specific scenes for an enlarged regression capacity.
Furthermore, we select distinct anchors before matching them with the ground
truth, which results in impressive scoring performance. Our approach enhances
all metrics compared to the baseline MTR, particularly with a notable relative
reduction of 13.5% in Miss Rate, resulting in state-of-the-art performance on
the Waymo Open Motion Dataset. Code is available at
https://github.com/Longzhong-Lin/EDA.
Related papers
- Adaptive Prediction Ensemble: Improving Out-of-Distribution Generalization of Motion Forecasting [15.916325272109454]
We propose a novel framework, Adaptive Prediction Ensemble (APE), which integrates deep learning and rule-based prediction experts.
A learned routing function, trained concurrently with the deep learning model, dynamically selects the most reliable prediction based on the input scenario.
This work highlights the potential of hybrid approaches for robust and generalizable motion prediction in autonomous driving.
arXiv Detail & Related papers (2024-07-12T17:57:00Z) - Ensemble Modeling for Multimodal Visual Action Recognition [50.38638300332429]
We propose an ensemble modeling approach for multimodal action recognition.
We independently train individual modality models using a variant of focal loss tailored to handle the long-tailed distribution of the MECCANO [21] dataset.
arXiv Detail & Related papers (2023-08-10T08:43:20Z) - ADAPT: Efficient Multi-Agent Trajectory Prediction with Adaptation [0.0]
ADAPT is a novel approach for jointly predicting the trajectories of all agents in the scene with dynamic weight learning.
Our approach outperforms state-of-the-art methods in both single-agent and multi-agent settings.
arXiv Detail & Related papers (2023-07-26T13:41:51Z) - MotionDiffuser: Controllable Multi-Agent Motion Prediction using
Diffusion [15.146808801331774]
MotionDiffuser is a diffusion based representation for the joint distribution of future trajectories over multiple agents.
We propose a general constrained sampling framework that enables controlled trajectory sampling based on differentiable cost functions.
We obtain state-of-the-art results for multi-agent motion prediction on the Open Motion dataset.
arXiv Detail & Related papers (2023-06-05T17:55:52Z) - Towards Consistent Stochastic Human Motion Prediction via Motion
Diffusion [8.10696589962658]
We propose DiffMotion as an end-to-end diffusion-based Human Motion Prediction framework.
Our results on benchmark datasets show that DiffMotion significantly outperforms previous methods in terms of both accuracy and fidelity.
arXiv Detail & Related papers (2023-05-21T19:31:56Z) - You Mostly Walk Alone: Analyzing Feature Attribution in Trajectory
Prediction [52.442129609979794]
Recent deep learning approaches for trajectory prediction show promising performance.
It remains unclear which features such black-box models actually learn to use for making predictions.
This paper proposes a procedure that quantifies the contributions of different cues to model performance.
arXiv Detail & Related papers (2021-10-11T14:24:15Z) - SMART: Simultaneous Multi-Agent Recurrent Trajectory Prediction [72.37440317774556]
We propose advances that address two key challenges in future trajectory prediction.
multimodality in both training data and predictions and constant time inference regardless of number of agents.
arXiv Detail & Related papers (2020-07-26T08:17:10Z) - Ambiguity in Sequential Data: Predicting Uncertain Futures with
Recurrent Models [110.82452096672182]
We propose an extension of the Multiple Hypothesis Prediction (MHP) model to handle ambiguous predictions with sequential data.
We also introduce a novel metric for ambiguous problems, which is better suited to account for uncertainties.
arXiv Detail & Related papers (2020-03-10T09:15:42Z) - Diversity inducing Information Bottleneck in Model Ensembles [73.80615604822435]
In this paper, we target the problem of generating effective ensembles of neural networks by encouraging diversity in prediction.
We explicitly optimize a diversity inducing adversarial loss for learning latent variables and thereby obtain diversity in the output predictions necessary for modeling multi-modal data.
Compared to the most competitive baselines, we show significant improvements in classification accuracy, under a shift in the data distribution.
arXiv Detail & Related papers (2020-03-10T03:10:41Z) - AvgOut: A Simple Output-Probability Measure to Eliminate Dull Responses [97.50616524350123]
We build dialogue models that are dynamically aware of what utterances or tokens are dull without any feature-engineering.
The first model, MinAvgOut, directly maximizes the diversity score through the output distributions of each batch.
The second model, Label Fine-Tuning (LFT), prepends to the source sequence a label continuously scaled by the diversity score to control the diversity level.
The third model, RL, adopts Reinforcement Learning and treats the diversity score as a reward signal.
arXiv Detail & Related papers (2020-01-15T18:32:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.