VRUNet: Multi-Task Learning Model for Intent Prediction of Vulnerable
Road Users
- URL: http://arxiv.org/abs/2007.05397v1
- Date: Fri, 10 Jul 2020 14:02:25 GMT
- Title: VRUNet: Multi-Task Learning Model for Intent Prediction of Vulnerable
Road Users
- Authors: Adithya Ranga, Filippo Giruzzi, Jagdish Bhanushali, Emilie Wirbel,
Patrick P\'erez, Tuan-Hung Vu and Xavier Perrotton
- Abstract summary: We propose a multi-task learning model to predict pedestrian actions, crossing intent and forecast their future path from video sequences.
We have trained the model on naturalistic driving open-source JAAD dataset, which is rich in behavioral annotations and real world scenarios.
Experimental results show state-of-the-art performance on JAAD dataset.
- Score: 3.6265173818019947
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Advanced perception and path planning are at the core for any self-driving
vehicle. Autonomous vehicles need to understand the scene and intentions of
other road users for safe motion planning. For urban use cases it is very
important to perceive and predict the intentions of pedestrians, cyclists,
scooters, etc., classified as vulnerable road users (VRU). Intent is a
combination of pedestrian activities and long term trajectories defining their
future motion. In this paper we propose a multi-task learning model to predict
pedestrian actions, crossing intent and forecast their future path from video
sequences. We have trained the model on naturalistic driving open-source JAAD
dataset, which is rich in behavioral annotations and real world scenarios.
Experimental results show state-of-the-art performance on JAAD dataset and how
we can benefit from jointly learning and predicting actions and trajectories
using 2D human pose features and scene context.
Related papers
- Local and Global Contextual Features Fusion for Pedestrian Intention
Prediction [2.203209457340481]
We analyse and analyse visual features of both pedestrian and traffic contexts.
To understand the global context, we utilise location, motion, and environmental information.
These multi-modality features are intelligently fused for effective intention learning.
arXiv Detail & Related papers (2023-05-01T22:37:31Z) - Safe Real-World Autonomous Driving by Learning to Predict and Plan with
a Mixture of Experts [3.2230833657560503]
We propose a distribution over multiple future trajectories for both the self-driving vehicle and other road agents.
During inference, we select the planning trajectory that minimizes a cost taking into account safety and the predicted probabilities.
We successfully deploy it on a self-driving vehicle on urban public roads, confirming that it drives safely without compromising comfort.
arXiv Detail & Related papers (2022-11-03T20:16:24Z) - Behavioral Intention Prediction in Driving Scenes: A Survey [70.53285924851767]
Behavioral Intention Prediction (BIP) simulates a human consideration process and fulfills the early prediction of specific behaviors.
This work provides a comprehensive review of BIP from the available datasets, key factors and challenges, pedestrian-centric and vehicle-centric BIP approaches, and BIP-aware applications.
arXiv Detail & Related papers (2022-11-01T11:07:37Z) - Pedestrian Stop and Go Forecasting with Hybrid Feature Fusion [87.77727495366702]
We introduce the new task of pedestrian stop and go forecasting.
Considering the lack of suitable existing datasets for it, we release TRANS, a benchmark for explicitly studying the stop and go behaviors of pedestrians in urban traffic.
We build it from several existing datasets annotated with pedestrians' walking motions, in order to have various scenarios and behaviors.
arXiv Detail & Related papers (2022-03-04T18:39:31Z) - Safety-aware Motion Prediction with Unseen Vehicles for Autonomous
Driving [104.32241082170044]
We study a new task, safety-aware motion prediction with unseen vehicles for autonomous driving.
Unlike the existing trajectory prediction task for seen vehicles, we aim at predicting an occupancy map.
Our approach is the first one that can predict the existence of unseen vehicles in most cases.
arXiv Detail & Related papers (2021-09-03T13:33:33Z) - Coupling Intent and Action for Pedestrian Crossing Behavior Prediction [25.54455403877285]
In this work, we follow the neuroscience and psychological literature to define pedestrian crossing behavior as a combination of an unobserved inner will and a set of multi-class actions.
We present a novel multi-task network that predicts future pedestrian actions and uses predicted future action as a prior to detect the present intent and action of the pedestrian.
arXiv Detail & Related papers (2021-05-10T06:26:25Z) - IntentNet: Learning to Predict Intention from Raw Sensor Data [86.74403297781039]
In this paper, we develop a one-stage detector and forecaster that exploits both 3D point clouds produced by a LiDAR sensor as well as dynamic maps of the environment.
Our multi-task model achieves better accuracy than the respective separate modules while saving computation, which is critical to reducing reaction time in self-driving applications.
arXiv Detail & Related papers (2021-01-20T00:31:52Z) - PePScenes: A Novel Dataset and Baseline for Pedestrian Action Prediction
in 3D [10.580548257913843]
We propose a new pedestrian action prediction dataset created by adding per-frame 2D/3D bounding box and behavioral annotations to nuScenes.
In addition, we propose a hybrid neural network architecture that incorporates various data modalities for predicting pedestrian crossing action.
arXiv Detail & Related papers (2020-12-14T18:13:44Z) - Pedestrian Intention Prediction: A Multi-task Perspective [83.7135926821794]
In order to be globally deployed, autonomous cars must guarantee the safety of pedestrians.
This work tries to solve this problem by jointly predicting the intention and visual states of pedestrians.
The method is a recurrent neural network in a multi-task learning approach.
arXiv Detail & Related papers (2020-10-20T13:42:31Z) - Spatiotemporal Relationship Reasoning for Pedestrian Intent Prediction [57.56466850377598]
Reasoning over visual data is a desirable capability for robotics and vision-based applications.
In this paper, we present a framework on graph to uncover relationships in different objects in the scene for reasoning about pedestrian intent.
Pedestrian intent, defined as the future action of crossing or not-crossing the street, is a very crucial piece of information for autonomous vehicles.
arXiv Detail & Related papers (2020-02-20T18:50:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.