Safety-Oriented Pedestrian Motion and Scene Occupancy Forecasting
- URL: http://arxiv.org/abs/2101.02385v1
- Date: Thu, 7 Jan 2021 06:08:21 GMT
- Title: Safety-Oriented Pedestrian Motion and Scene Occupancy Forecasting
- Authors: Katie Luo, Sergio Casas, Renjie Liao, Xinchen Yan, Yuwen Xiong,
Wenyuan Zeng, Raquel Urtasun
- Abstract summary: We advocate for predicting both the individual motions as well as the scene occupancy map.
We propose a Scene-Actor Graph Neural Network (SA-GNN) which preserves the relative spatial information of pedestrians.
On two large-scale real-world datasets, we showcase that our scene-occupancy predictions are more accurate and better calibrated than those from state-of-the-art motion forecasting methods.
- Score: 91.69900691029908
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we address the important problem in self-driving of
forecasting multi-pedestrian motion and their shared scene occupancy map,
critical for safe navigation. Our contributions are two-fold. First, we
advocate for predicting both the individual motions as well as the scene
occupancy map in order to effectively deal with missing detections caused by
postprocessing, e.g., confidence thresholding and non-maximum suppression.
Second, we propose a Scene-Actor Graph Neural Network (SA-GNN) which preserves
the relative spatial information of pedestrians via 2D convolution, and
captures the interactions among pedestrians within the same scene, including
those that have not been detected, via message passing. On two large-scale
real-world datasets, nuScenes and ATG4D, we showcase that our scene-occupancy
predictions are more accurate and better calibrated than those from
state-of-the-art motion forecasting methods, while also matching their
performance in pedestrian motion forecasting metrics.
Related papers
- PIP-Net: Pedestrian Intention Prediction in the Wild [11.799731429829603]
PIP-Net is a novel framework designed to predict pedestrian crossing intentions by AVs in real-world urban scenarios.
We offer two variants of PIP-Net designed for different camera mounts and setups.
The proposed model employs a recurrent and temporal attention-based solution, outperforming state-of-the-art performance.
For the first time, we present the Urban-PIP dataset, a customised pedestrian intention prediction dataset.
arXiv Detail & Related papers (2024-02-20T08:28:45Z) - Implicit Occupancy Flow Fields for Perception and Prediction in
Self-Driving [68.95178518732965]
A self-driving vehicle (SDV) must be able to perceive its surroundings and predict the future behavior of other traffic participants.
Existing works either perform object detection followed by trajectory of the detected objects, or predict dense occupancy and flow grids for the whole scene.
This motivates our unified approach to perception and future prediction that implicitly represents occupancy and flow over time with a single neural network.
arXiv Detail & Related papers (2023-08-02T23:39:24Z) - SGCN:Sparse Graph Convolution Network for Pedestrian Trajectory
Prediction [64.16212996247943]
We present a Sparse Graph Convolution Network(SGCN) for pedestrian trajectory prediction.
Specifically, the SGCN explicitly models the sparse directed interaction with a sparse directed spatial graph to capture adaptive interaction pedestrians.
visualizations indicate that our method can capture adaptive interactions between pedestrians and their effective motion tendencies.
arXiv Detail & Related papers (2021-04-04T03:17:42Z) - Pedestrian Motion State Estimation From 2D Pose [3.189006905282788]
Traffic violation and the flexible and changeable nature of pedestrians make it more difficult to predict pedestrian behavior or intention.
In combination with pedestrian motion state and other influencing factors, pedestrian intention can be predicted to avoid unnecessary accidents.
This paper verifies the proposed algorithm on the JAAD public dataset, and the accuracy is improved by 11.6% compared with the existing method.
arXiv Detail & Related papers (2021-02-27T07:00:06Z) - A Real-Time Predictive Pedestrian Collision Warning Service for
Cooperative Intelligent Transportation Systems Using 3D Pose Estimation [10.652350454373531]
We propose a real-time predictive pedestrian collision warning service (P2CWS) for two tasks: pedestrian orientation recognition (100.53 FPS) and intention prediction (35.76 FPS)
Our framework obtains satisfying generalization over multiple sites because of the proposed site-independent features.
The proposed vision framework realizes 89.3% accuracy in the behavior recognition task on the TUD dataset without any training process.
arXiv Detail & Related papers (2020-09-23T00:55:12Z) - Recognition and 3D Localization of Pedestrian Actions from Monocular
Video [11.29865843123467]
This paper focuses on monocular pedestrian action recognition and 3D localization from an egocentric view.
A challenge in addressing this problem in urban traffic scenes is attributed to the unpredictable behavior of pedestrians.
arXiv Detail & Related papers (2020-08-03T19:57:03Z) - Implicit Latent Variable Model for Scene-Consistent Motion Forecasting [78.74510891099395]
In this paper, we aim to learn scene-consistent motion forecasts of complex urban traffic directly from sensor data.
We model the scene as an interaction graph and employ powerful graph neural networks to learn a distributed latent representation of the scene.
arXiv Detail & Related papers (2020-07-23T14:31:25Z) - A Spatial-Temporal Attentive Network with Spatial Continuity for
Trajectory Prediction [74.00750936752418]
We propose a novel model named spatial-temporal attentive network with spatial continuity (STAN-SC)
First, spatial-temporal attention mechanism is presented to explore the most useful and important information.
Second, we conduct a joint feature sequence based on the sequence and instant state information to make the generative trajectories keep spatial continuity.
arXiv Detail & Related papers (2020-03-13T04:35:50Z) - Spatiotemporal Relationship Reasoning for Pedestrian Intent Prediction [57.56466850377598]
Reasoning over visual data is a desirable capability for robotics and vision-based applications.
In this paper, we present a framework on graph to uncover relationships in different objects in the scene for reasoning about pedestrian intent.
Pedestrian intent, defined as the future action of crossing or not-crossing the street, is a very crucial piece of information for autonomous vehicles.
arXiv Detail & Related papers (2020-02-20T18:50:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.