3D Scene Understanding at Urban Intersection using Stereo Vision and
Digital Map
- URL: http://arxiv.org/abs/2112.05295v1
- Date: Fri, 10 Dec 2021 02:05:15 GMT
- Title: 3D Scene Understanding at Urban Intersection using Stereo Vision and
Digital Map
- Authors: Prarthana Bhattacharyya, Yanlei Gu, Jiali Bao, Xu Liu and Shunsuke
Kamijo
- Abstract summary: We introduce a stereo vision and 3D digital map based approach to spatially and temporally analyze the traffic situation at urban intersections.
We qualitatively and quantitatively evaluate our proposed technique on real traffic data collected at an urban canyon in Tokyo to demonstrate the efficacy of the system.
- Score: 4.640144833676576
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The driving behavior at urban intersections is very complex. It is thus
crucial for autonomous vehicles to comprehensively understand challenging urban
traffic scenes in order to navigate intersections and prevent accidents. In
this paper, we introduce a stereo vision and 3D digital map based approach to
spatially and temporally analyze the traffic situation at urban intersections.
Stereo vision is used to detect, classify and track obstacles, while a 3D
digital map is used to improve ego-localization and provide context in terms of
road-layout information. A probabilistic approach that temporally integrates
these geometric, semantic, dynamic and contextual cues is presented. We
qualitatively and quantitatively evaluate our proposed technique on real
traffic data collected at an urban canyon in Tokyo to demonstrate the efficacy
of the system in providing comprehensive awareness of the traffic surroundings.
Related papers
- Neural Semantic Map-Learning for Autonomous Vehicles [85.8425492858912]
We present a mapping system that fuses local submaps gathered from a fleet of vehicles at a central instance to produce a coherent map of the road environment.
Our method jointly aligns and merges the noisy and incomplete local submaps using a scene-specific Neural Signed Distance Field.
We leverage memory-efficient sparse feature-grids to scale to large areas and introduce a confidence score to model uncertainty in scene reconstruction.
arXiv Detail & Related papers (2024-10-10T10:10:03Z) - Empowering Urban Traffic Management: Elevated 3D LiDAR for Data Collection and Advanced Object Detection Analysis [4.831084635928491]
This paper presents a novel framework that transforms the detection and analysis of 3D objects in traffic scenarios by utilizing the power of elevated LiDAR sensors.
Due to the limitation in obtaining real-world traffic datasets, we utilize the simulator to generate 3D point cloud for specific scenarios.
arXiv Detail & Related papers (2024-05-21T21:12:09Z) - Graph Attention Network for Lane-Wise and Topology-Invariant Intersection Traffic Simulation [8.600701437207725]
We propose two efficient and accurate "Digital Twin" models for intersections.
These digital twins capture temporal, spatial, and contextual aspects of traffic within intersections.
Our study's applications extend to lane reconfiguration, driving behavior analysis, and facilitating informed decisions regarding intersection safety and efficiency enhancements.
arXiv Detail & Related papers (2024-04-11T03:02:06Z) - OpenLane-V2: A Topology Reasoning Benchmark for Unified 3D HD Mapping [84.65114565766596]
We present OpenLane-V2, the first dataset on topology reasoning for traffic scene structure.
OpenLane-V2 consists of 2,000 annotated road scenes that describe traffic elements and their correlation to the lanes.
We evaluate various state-of-the-art methods, and present their quantitative and qualitative results on OpenLane-V2 to indicate future avenues for investigating topology reasoning in traffic scenes.
arXiv Detail & Related papers (2023-04-20T16:31:22Z) - Street-View Image Generation from a Bird's-Eye View Layout [95.36869800896335]
Bird's-Eye View (BEV) Perception has received increasing attention in recent years.
Data-driven simulation for autonomous driving has been a focal point of recent research.
We propose BEVGen, a conditional generative model that synthesizes realistic and spatially consistent surrounding images.
arXiv Detail & Related papers (2023-01-11T18:39:34Z) - Traffic-Net: 3D Traffic Monitoring Using a Single Camera [1.1602089225841632]
We provide a practical platform for real-time traffic monitoring using a single CCTV traffic camera.
We adapt a custom YOLOv5 deep neural network model for vehicle/pedestrian detection and an enhanced SORT tracking algorithm.
We also develop a hierarchical traffic modelling solution based on short- and long-term temporal video data stream.
arXiv Detail & Related papers (2021-09-19T16:59:01Z) - Evaluating Computer Vision Techniques for Urban Mobility on Large-Scale,
Unconstrained Roads [25.29906312974705]
This paper proposes a simple mobile imaging setup to address several common problems in road safety at scale.
We use recent computer vision techniques to identify possible irregularities on roads.
We also demonstrate the mobile imaging solution's applicability to spot traffic violations.
arXiv Detail & Related papers (2021-09-11T09:07:56Z) - An Experimental Urban Case Study with Various Data Sources and a Model
for Traffic Estimation [65.28133251370055]
We organize an experimental campaign with video measurement in an area within the urban network of Zurich, Switzerland.
We focus on capturing the traffic state in terms of traffic flow and travel times by ensuring measurements from established thermal cameras.
We propose a simple yet efficient Multiple Linear Regression (MLR) model to estimate travel times with fusion of various data sources.
arXiv Detail & Related papers (2021-08-02T08:13:57Z) - Multi-Modal Fusion Transformer for End-to-End Autonomous Driving [59.60483620730437]
We propose TransFuser, a novel Multi-Modal Fusion Transformer, to integrate image and LiDAR representations using attention.
Our approach achieves state-of-the-art driving performance while reducing collisions by 76% compared to geometry-based fusion.
arXiv Detail & Related papers (2021-04-19T11:48:13Z) - Fine-Grained Vehicle Perception via 3D Part-Guided Visual Data
Augmentation [77.60050239225086]
We propose an effective training data generation process by fitting a 3D car model with dynamic parts to vehicles in real images.
Our approach is fully automatic without any human interaction.
We present a multi-task network for VUS parsing and a multi-stream network for VHI parsing.
arXiv Detail & Related papers (2020-12-15T03:03:38Z) - Recognition and 3D Localization of Pedestrian Actions from Monocular
Video [11.29865843123467]
This paper focuses on monocular pedestrian action recognition and 3D localization from an egocentric view.
A challenge in addressing this problem in urban traffic scenes is attributed to the unpredictable behavior of pedestrians.
arXiv Detail & Related papers (2020-08-03T19:57:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.