MS-Net: A Multi-Path Sparse Model for Motion Prediction in Multi-Scenes
- URL: http://arxiv.org/abs/2403.00353v1
- Date: Fri, 1 Mar 2024 08:32:12 GMT
- Title: MS-Net: A Multi-Path Sparse Model for Motion Prediction in Multi-Scenes
- Authors: Xiaqiang Tang, Weigao Sun, Siyuan Hu, Yiyang Sun, Yafeng Guo
- Abstract summary: Multi-Scenes Network (aka MS-Net) is a multi-path sparse model trained by an evolutionary process.
MS-Net selectively activates a subset of its parameters during the inference stage to produce prediction results for each scene.
Our experiment results show that MS-Net outperforms existing state-of-the-art methods on well-established pedestrian motion prediction datasets.
- Score: 1.4451387915783602
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The multi-modality and stochastic characteristics of human behavior make
motion prediction a highly challenging task, which is critical for autonomous
driving. While deep learning approaches have demonstrated their great potential
in this area, it still remains unsolved to establish a connection between
multiple driving scenes (e.g., merging, roundabout, intersection) and the
design of deep learning models. Current learning-based methods typically use
one unified model to predict trajectories in different scenarios, which may
result in sub-optimal results for one individual scene. To address this issue,
we propose Multi-Scenes Network (aka. MS-Net), which is a multi-path sparse
model trained by an evolutionary process. MS-Net selectively activates a subset
of its parameters during the inference stage to produce prediction results for
each scene. In the training stage, the motion prediction task under
differentiated scenes is abstracted as a multi-task learning problem, an
evolutionary algorithm is designed to encourage the network search of the
optimal parameters for each scene while sharing common knowledge between
different scenes. Our experiment results show that with substantially reduced
parameters, MS-Net outperforms existing state-of-the-art methods on
well-established pedestrian motion prediction datasets, e.g., ETH and UCY, and
ranks the 2nd place on the INTERACTION challenge.
Related papers
- Multi-Transmotion: Pre-trained Model for Human Motion Prediction [68.87010221355223]
Multi-Transmotion is an innovative transformer-based model designed for cross-modality pre-training.
Our methodology demonstrates competitive performance across various datasets on several downstream tasks.
arXiv Detail & Related papers (2024-11-04T23:15:21Z) - Joint Depth Prediction and Semantic Segmentation with Multi-View SAM [59.99496827912684]
We propose a Multi-View Stereo (MVS) technique for depth prediction that benefits from rich semantic features of the Segment Anything Model (SAM)
This enhanced depth prediction, in turn, serves as a prompt to our Transformer-based semantic segmentation decoder.
arXiv Detail & Related papers (2023-10-31T20:15:40Z) - EUCLID: Towards Efficient Unsupervised Reinforcement Learning with
Multi-choice Dynamics Model [46.99510778097286]
Unsupervised reinforcement learning (URL) poses a promising paradigm to learn useful behaviors in a task-agnostic environment.
We introduce a novel model-fused paradigm to jointly pre-train the dynamics model and unsupervised exploration policy in the pre-training phase.
We show that EUCLID achieves state-of-the-art performance with high sample efficiency.
arXiv Detail & Related papers (2022-10-02T12:11:44Z) - Instance-Aware Predictive Navigation in Multi-Agent Environments [93.15055834395304]
We propose an Instance-Aware Predictive Control (IPC) approach, which forecasts interactions between agents as well as future scene structures.
We adopt a novel multi-instance event prediction module to estimate the possible interaction among agents in the ego-centric view.
We design a sequential action sampling strategy to better leverage predicted states on both scene-level and instance-level.
arXiv Detail & Related papers (2021-01-14T22:21:25Z) - Pedestrian Behavior Prediction via Multitask Learning and Categorical
Interaction Modeling [13.936894582450734]
We propose a multitask learning framework that simultaneously predicts trajectories and actions of pedestrians by relying on multimodal data.
We show that our model achieves state-of-the-art performance and improves trajectory and action prediction by up to 22% and 6% respectively.
arXiv Detail & Related papers (2020-12-06T15:57:11Z) - Trajectory-wise Multiple Choice Learning for Dynamics Generalization in
Reinforcement Learning [137.39196753245105]
We present a new model-based reinforcement learning algorithm that learns a multi-headed dynamics model for dynamics generalization.
We incorporate context learning, which encodes dynamics-specific information from past experiences into the context latent vector.
Our method exhibits superior zero-shot generalization performance across a variety of control tasks, compared to state-of-the-art RL methods.
arXiv Detail & Related papers (2020-10-26T03:20:42Z) - SMART: Simultaneous Multi-Agent Recurrent Trajectory Prediction [72.37440317774556]
We propose advances that address two key challenges in future trajectory prediction.
multimodality in both training data and predictions and constant time inference regardless of number of agents.
arXiv Detail & Related papers (2020-07-26T08:17:10Z) - MultiXNet: Multiclass Multistage Multimodal Motion Prediction [27.046311751308775]
MultiXNet is an end-to-end approach for detection and motion prediction based directly on lidar sensor data.
The method was evaluated on large-scale, real-world data collected by a fleet of SDVs in several cities.
arXiv Detail & Related papers (2020-06-03T01:01:48Z) - Multi-Task Learning for Dense Prediction Tasks: A Survey [87.66280582034838]
Multi-task learning (MTL) techniques have shown promising results w.r.t. performance, computations and/or memory footprint.
We provide a well-rounded view on state-of-the-art deep learning approaches for MTL in computer vision.
arXiv Detail & Related papers (2020-04-28T09:15:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.