Pedestrian Environment Model for Automated Driving
- URL: http://arxiv.org/abs/2308.09080v1
- Date: Thu, 17 Aug 2023 16:10:58 GMT
- Title: Pedestrian Environment Model for Automated Driving
- Authors: Adrian Holzbock, Alexander Tsaregorodtsev, and Vasileios Belagiannis
- Abstract summary: We propose an environment model that includes the position of the pedestrians as well as their pose information.
We extract the skeletal information with a neural network human pose estimator from the image.
To obtain the 3D information of the position, we aggregate the data from consecutive frames in conjunction with the vehicle position.
- Score: 54.16257759472116
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Besides interacting correctly with other vehicles, automated vehicles should
also be able to react in a safe manner to vulnerable road users like
pedestrians or cyclists. For a safe interaction between pedestrians and
automated vehicles, the vehicle must be able to interpret the pedestrian's
behavior. Common environment models do not contain information like body poses
used to understand the pedestrian's intent. In this work, we propose an
environment model that includes the position of the pedestrians as well as
their pose information. We only use images from a monocular camera and the
vehicle's localization data as input to our pedestrian environment model. We
extract the skeletal information with a neural network human pose estimator
from the image. Furthermore, we track the skeletons with a simple tracking
algorithm based on the Hungarian algorithm and an ego-motion compensation. To
obtain the 3D information of the position, we aggregate the data from
consecutive frames in conjunction with the vehicle position. We demonstrate our
pedestrian environment model on data generated with the CARLA simulator and the
nuScenes dataset. Overall, we reach a relative position error of around 16% on
both datasets.
Related papers
- ViFi-Loc: Multi-modal Pedestrian Localization using GAN with
Camera-Phone Correspondences [7.953401800573514]
We propose a Generative Adversarial Network architecture to produce more accurate location estimations for pedestrians.
During training, it learns the underlying linkage between pedestrians' camera-phone data correspondences.
We show that our GAN produces 3D coordinates at 1 to 2 meter localization error across 5 different outdoor scenes.
arXiv Detail & Related papers (2022-11-22T05:27:38Z) - Decentralized Vehicle Coordination: The Berkeley DeepDrive Drone Dataset [103.35624417260541]
Decentralized vehicle coordination is useful in understructured road environments.
We collect the Berkeley DeepDrive Drone dataset to study implicit "social etiquette" observed by nearby drivers.
The dataset is of primary interest for studying decentralized multiagent planning employed by human drivers and for computer vision in remote sensing settings.
arXiv Detail & Related papers (2022-09-19T05:06:57Z) - WatchPed: Pedestrian Crossing Intention Prediction Using Embedded
Sensors of Smartwatch [11.83842808044211]
We design, implement, and evaluate the first pedestrian intention prediction model based on integration of motion sensor data.
A novel machine learning architecture is proposed to effectively incorporate the motion sensor data.
We present the first pedestrian intention prediction dataset integrated with time-synchronized motion sensor data.
arXiv Detail & Related papers (2022-08-15T21:27:21Z) - Euro-PVI: Pedestrian Vehicle Interactions in Dense Urban Centers [126.81938540470847]
We propose Euro-PVI, a dataset of pedestrian and bicyclist trajectories.
In this work, we develop a joint inference model that learns an expressive multi-modal shared latent space across agents in the urban scene.
We achieve state of the art results on the nuScenes and Euro-PVI datasets demonstrating the importance of capturing interactions between ego-vehicle and pedestrians (bicyclists) for accurate predictions.
arXiv Detail & Related papers (2021-06-22T15:40:21Z) - Fine-Grained Vehicle Perception via 3D Part-Guided Visual Data
Augmentation [77.60050239225086]
We propose an effective training data generation process by fitting a 3D car model with dynamic parts to vehicles in real images.
Our approach is fully automatic without any human interaction.
We present a multi-task network for VUS parsing and a multi-stream network for VHI parsing.
arXiv Detail & Related papers (2020-12-15T03:03:38Z) - RAIST: Learning Risk Aware Traffic Interactions via Spatio-Temporal
Graph Convolutional Networks [19.582873794287632]
A key aspect of driving a road vehicle is to interact with other road users, assess their intentions and make risk-aware tactical decisions.
We propose a novel driving framework for egocentric views based on traffic graphs.
We claim that our framework learns risk-aware representations by improving on the task of risk object identification.
arXiv Detail & Related papers (2020-11-17T15:49:22Z) - Recovering and Simulating Pedestrians in the Wild [81.38135735146015]
We propose to recover the shape and motion of pedestrians from sensor readings captured in the wild by a self-driving car driving around.
We incorporate the reconstructed pedestrian assets bank in a realistic 3D simulation system.
We show that the simulated LiDAR data can be used to significantly reduce the amount of real-world data required for visual perception tasks.
arXiv Detail & Related papers (2020-11-16T17:16:32Z) - Hidden Footprints: Learning Contextual Walkability from 3D Human Trails [70.01257397390361]
Current datasets only tell you where people are, not where they could be.
We first augment the set of valid, labeled walkable regions by propagating person observations between images, utilizing 3D information to create what we call hidden footprints.
We devise a training strategy designed for such sparse labels, combining a class-balanced classification loss with a contextual adversarial loss.
arXiv Detail & Related papers (2020-08-19T23:19:08Z) - SurfelGAN: Synthesizing Realistic Sensor Data for Autonomous Driving [27.948417322786575]
We present a simple yet effective approach to generate realistic scenario sensor data.
Our approach uses texture-mapped surfels to efficiently reconstruct the scene from an initial vehicle pass or set of passes.
We then leverage a SurfelGAN network to reconstruct realistic camera images for novel positions and orientations of the self-driving vehicle.
arXiv Detail & Related papers (2020-05-08T04:01:14Z) - The Pedestrian Patterns Dataset [11.193504036335503]
The dataset was collected by repeatedly traversing the same three routes for one week starting at different specific timeslots.
The purpose of the dataset is to capture the patterns of social and pedestrian behavior along the traversed routes at different times.
arXiv Detail & Related papers (2020-01-06T23:58:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.