Interaction Detection Between Vehicles and Vulnerable Road Users: A Deep
Generative Approach with Attention
- URL: http://arxiv.org/abs/2105.03891v1
- Date: Sun, 9 May 2021 10:03:55 GMT
- Title: Interaction Detection Between Vehicles and Vulnerable Road Users: A Deep
Generative Approach with Attention
- Authors: Hao Cheng, Li Feng, Hailong Liu, Takatsugu Hirayama, Hiroshi Murase
and Monika Sester
- Abstract summary: We propose a conditional generative model for interaction detection at intersections.
It aims to automatically analyze massive video data about the continuity of road users' behavior.
The model's efficacy was validated by testing on real-world datasets.
- Score: 9.442285577226606
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Intersections where vehicles are permitted to turn and interact with
vulnerable road users (VRUs) like pedestrians and cyclists are among some of
the most challenging locations for automated and accurate recognition of road
users' behavior. In this paper, we propose a deep conditional generative model
for interaction detection at such locations. It aims to automatically analyze
massive video data about the continuity of road users' behavior. This task is
essential for many intelligent transportation systems such as traffic safety
control and self-driving cars that depend on the understanding of road users'
locomotion. A Conditional Variational Auto-Encoder based model with Gaussian
latent variables is trained to encode road users' behavior and perform
probabilistic and diverse predictions of interactions. The model takes as input
the information of road users' type, position and motion automatically
extracted by a deep learning object detector and optical flow from videos, and
generates frame-wise probabilities that represent the dynamics of interactions
between a turning vehicle and any VRUs involved. The model's efficacy was
validated by testing on real--world datasets acquired from two different
intersections. It achieved an F1-score above 0.96 at a right--turn intersection
in Germany and 0.89 at a left--turn intersection in Japan, both with very busy
traffic flows.
Related papers
- Identifying built environment factors influencing driver yielding behavior at unsignalized intersections: A naturalistic open-source dataset collected in Minnesota [1.7810134788247751]
Many factors influence the yielding result of a driver-pedestrian interaction, including traffic volume, vehicle speed, roadway characteristics, etc.
Our study introduces an extensive open-source dataset, compiled from video data at 18 unsignalized intersections across Minnesota.
Documenting more than 3000 interactions, this dataset provides a detailed view of driver-pedestrian interactions and over 50 distinct contextual variables.
arXiv Detail & Related papers (2023-12-22T23:18:27Z) - Learning Driver Models for Automated Vehicles via Knowledge Sharing and
Personalization [2.07180164747172]
This paper describes a framework for learning Automated Vehicles (AVs) driver models via knowledge sharing between vehicles and personalization.
It finds several applications across transportation engineering including intelligent transportation systems, traffic management, and vehicle-to-vehicle communication.
arXiv Detail & Related papers (2023-08-31T17:18:15Z) - Exploring Contextual Representation and Multi-Modality for End-to-End
Autonomous Driving [58.879758550901364]
Recent perception systems enhance spatial understanding with sensor fusion but often lack full environmental context.
We introduce a framework that integrates three cameras to emulate the human field of view, coupled with top-down bird-eye-view semantic data to enhance contextual representation.
Our method achieves displacement error by 0.67m in open-loop settings, surpassing current methods by 6.9% on the nuScenes dataset.
arXiv Detail & Related papers (2022-10-13T05:56:20Z) - Learning energy-efficient driving behaviors by imitating experts [75.12960180185105]
This paper examines the role of imitation learning in bridging the gap between control strategies and realistic limitations in communication and sensing.
We show that imitation learning can succeed in deriving policies that, if adopted by 5% of vehicles, may boost the energy-efficiency of networks with varying traffic conditions by 15% using only local observations.
arXiv Detail & Related papers (2022-06-28T17:08:31Z) - Pedestrian Trajectory Prediction via Spatial Interaction Transformer
Network [7.150832716115448]
In traffic scenes, when encountering with oncoming people, pedestrians may make sudden turns or stop immediately.
To predict such unpredictable trajectories, we can gain insights into the interaction between pedestrians.
We present a novel generative method named Spatial Interaction Transformer (SIT), which learns the correlation of pedestrian trajectories through attention mechanisms.
arXiv Detail & Related papers (2021-12-13T13:08:04Z) - Automated Object Behavioral Feature Extraction for Potential Risk
Analysis based on Video Sensor [6.291501119156943]
Pedestrians are exposed to risk of death or serious injuries on roads, especially unsignalized crosswalks.
We propose an automated and simpler system for effectively extracting object behavioral features from video sensors deployed on the road.
This study demonstrates the potential for a network of connected video sensors to provide actionable data for smart cities.
arXiv Detail & Related papers (2021-07-08T01:11:31Z) - Euro-PVI: Pedestrian Vehicle Interactions in Dense Urban Centers [126.81938540470847]
We propose Euro-PVI, a dataset of pedestrian and bicyclist trajectories.
In this work, we develop a joint inference model that learns an expressive multi-modal shared latent space across agents in the urban scene.
We achieve state of the art results on the nuScenes and Euro-PVI datasets demonstrating the importance of capturing interactions between ego-vehicle and pedestrians (bicyclists) for accurate predictions.
arXiv Detail & Related papers (2021-06-22T15:40:21Z) - End-to-End Intersection Handling using Multi-Agent Deep Reinforcement
Learning [63.56464608571663]
Navigating through intersections is one of the main challenging tasks for an autonomous vehicle.
In this work, we focus on the implementation of a system able to navigate through intersections where only traffic signs are provided.
We propose a multi-agent system using a continuous, model-free Deep Reinforcement Learning algorithm used to train a neural network for predicting both the acceleration and the steering angle at each time step.
arXiv Detail & Related papers (2021-04-28T07:54:40Z) - Fine-Grained Vehicle Perception via 3D Part-Guided Visual Data
Augmentation [77.60050239225086]
We propose an effective training data generation process by fitting a 3D car model with dynamic parts to vehicles in real images.
Our approach is fully automatic without any human interaction.
We present a multi-task network for VUS parsing and a multi-stream network for VHI parsing.
arXiv Detail & Related papers (2020-12-15T03:03:38Z) - Detecting 32 Pedestrian Attributes for Autonomous Vehicles [103.87351701138554]
In this paper, we address the problem of jointly detecting pedestrians and recognizing 32 pedestrian attributes.
We introduce a Multi-Task Learning (MTL) model relying on a composite field framework, which achieves both goals in an efficient way.
We show competitive detection and attribute recognition results, as well as a more stable MTL training.
arXiv Detail & Related papers (2020-12-04T15:10:12Z) - RAIST: Learning Risk Aware Traffic Interactions via Spatio-Temporal
Graph Convolutional Networks [19.582873794287632]
A key aspect of driving a road vehicle is to interact with other road users, assess their intentions and make risk-aware tactical decisions.
We propose a novel driving framework for egocentric views based on traffic graphs.
We claim that our framework learns risk-aware representations by improving on the task of risk object identification.
arXiv Detail & Related papers (2020-11-17T15:49:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.