Pedestrian Intention Prediction: A Multi-task Perspective
- URL: http://arxiv.org/abs/2010.10270v2
- Date: Thu, 20 May 2021 11:14:35 GMT
- Title: Pedestrian Intention Prediction: A Multi-task Perspective
- Authors: Smail Ait Bouhsain, Saeed Saadatnejad and Alexandre Alahi
- Abstract summary: In order to be globally deployed, autonomous cars must guarantee the safety of pedestrians.
This work tries to solve this problem by jointly predicting the intention and visual states of pedestrians.
The method is a recurrent neural network in a multi-task learning approach.
- Score: 83.7135926821794
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In order to be globally deployed, autonomous cars must guarantee the safety
of pedestrians. This is the reason why forecasting pedestrians' intentions
sufficiently in advance is one of the most critical and challenging tasks for
autonomous vehicles. This work tries to solve this problem by jointly
predicting the intention and visual states of pedestrians. In terms of visual
states, whereas previous work focused on x-y coordinates, we will also predict
the size and indeed the whole bounding box of the pedestrian. The method is a
recurrent neural network in a multi-task learning approach. It has one head
that predicts the intention of the pedestrian for each one of its future
position and another one predicting the visual states of the pedestrian.
Experiments on the JAAD dataset show the superiority of the performance of our
method compared to previous works for intention prediction. Also, although its
simple architecture (more than 2 times faster), the performance of the bounding
box prediction is comparable to the ones yielded by much more complex
architectures. Our code is available online.
Related papers
- Pedestrian 3D Bounding Box Prediction [83.7135926821794]
We focus on 3D bounding boxes, which are reasonable estimates of humans without modeling complex motion details for autonomous vehicles.
We suggest this new problem and present a simple yet effective model for pedestrians' 3D bounding box prediction.
This method follows an encoder-decoder architecture based on recurrent neural networks.
arXiv Detail & Related papers (2022-06-28T17:59:45Z) - DR.VIC: Decomposition and Reasoning for Video Individual Counting [93.12166351940242]
We propose to conduct pedestrian counting from a new perspective - Video Individual Counting (VIC)
Instead of relying on the Multiple Object Tracking (MOT) techniques, we propose to solve the problem by decomposing all pedestrians into the initial pedestrians who existed in the first frame and the new pedestrians with separate identities in each following frame.
An end-to-end Decomposition and Reasoning Network (DRNet) is designed to predict the initial pedestrian count with the density estimation method and reason the new pedestrian's count of each frame with the differentiable optimal transport.
arXiv Detail & Related papers (2022-03-23T11:24:44Z) - Pedestrian Stop and Go Forecasting with Hybrid Feature Fusion [87.77727495366702]
We introduce the new task of pedestrian stop and go forecasting.
Considering the lack of suitable existing datasets for it, we release TRANS, a benchmark for explicitly studying the stop and go behaviors of pedestrians in urban traffic.
We build it from several existing datasets annotated with pedestrians' walking motions, in order to have various scenarios and behaviors.
arXiv Detail & Related papers (2022-03-04T18:39:31Z) - "If you could see me through my eyes": Predicting Pedestrian Perception [0.0]
We use synthetic data from simulations of aspecific pedestrian crossing scenario to train avariational autoencoder and along short-term memory network.
We can accurately predict apedestrian's future perceptions within relevant time horizons.
Such trained networks can later be used to predict pedestrian behaviors even from the perspective of the autonomous car.
arXiv Detail & Related papers (2022-02-28T17:36:12Z) - PSI: A Pedestrian Behavior Dataset for Socially Intelligent Autonomous
Car [47.01116716025731]
This paper proposes and shares another benchmark dataset called the IUPUI-CSRC Pedestrian Situated Intent (PSI) data.
The first novel label is the dynamic intent changes for the pedestrians to cross in front of the ego-vehicle, achieved from 24 drivers.
The second one is the text-based explanations of the driver reasoning process when estimating pedestrian intents and predicting their behaviors.
arXiv Detail & Related papers (2021-12-05T15:54:57Z) - Single Shot Multitask Pedestrian Detection and Behavior Prediction [9.147707153504117]
We propose a novel architecture using spatial-temporal multi-tasking to do camera based pedestrian detection and intention prediction.
Our approach significantly reduces the latency by being able to detect and predict all pedestrians' intention in a single shot manner.
arXiv Detail & Related papers (2021-01-06T19:10:23Z) - Pedestrian Behavior Prediction for Automated Driving: Requirements,
Metrics, and Relevant Features [1.1888947789336193]
We analyze the requirements on pedestrian behavior prediction for automated driving via a system-level approach.
Based on human driving behavior we derive appropriate reaction patterns of an automated vehicle.
We present a pedestrian prediction model based on a Variational Conditional Auto-Encoder which incorporates multiple contextual cues.
arXiv Detail & Related papers (2020-12-15T16:52:49Z) - Spatiotemporal Relationship Reasoning for Pedestrian Intent Prediction [57.56466850377598]
Reasoning over visual data is a desirable capability for robotics and vision-based applications.
In this paper, we present a framework on graph to uncover relationships in different objects in the scene for reasoning about pedestrian intent.
Pedestrian intent, defined as the future action of crossing or not-crossing the street, is a very crucial piece of information for autonomous vehicles.
arXiv Detail & Related papers (2020-02-20T18:50:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.