MC-BEVRO: Multi-Camera Bird Eye View Road Occupancy Detection for Traffic Monitoring
- URL: http://arxiv.org/abs/2502.11287v1
- Date: Sun, 16 Feb 2025 22:03:03 GMT
- Title: MC-BEVRO: Multi-Camera Bird Eye View Road Occupancy Detection for Traffic Monitoring
- Authors: Arpitsinh Vaghela, Duo Lu, Aayush Atul Verma, Bharatesh Chakravarthi, Hua Wei, Yezhou Yang,
- Abstract summary: Single camera 3D perception for traffic monitoring faces significant challenges due to occlusion and limited field of view.
This paper introduces a novel Bird's-Eye-View road occupancy detection framework that leverages multiple roadside cameras.
- Score: 23.396192711865147
- License:
- Abstract: Single camera 3D perception for traffic monitoring faces significant challenges due to occlusion and limited field of view. Moreover, fusing information from multiple cameras at the image feature level is difficult because of different view angles. Further, the necessity for practical implementation and compatibility with existing traffic infrastructure compounds these challenges. To address these issues, this paper introduces a novel Bird's-Eye-View road occupancy detection framework that leverages multiple roadside cameras to overcome the aforementioned limitations. To facilitate the framework's development and evaluation, a synthetic dataset featuring diverse scenes and varying camera configurations is generated using the CARLA simulator. A late fusion and three early fusion methods were implemented within the proposed framework, with performance further enhanced by integrating backgrounds. Extensive evaluations were conducted to analyze the impact of multi-camera inputs and varying BEV occupancy map sizes on model performance. Additionally, a real-world data collection pipeline was developed to assess the model's ability to generalize to real-world environments. The sim-to-real capabilities of the model were evaluated using zero-shot and few-shot fine-tuning, demonstrating its potential for practical application. This research aims to advance perception systems in traffic monitoring, contributing to improved traffic management, operational efficiency, and road safety.
Related papers
- TopView: Vectorising road users in a bird's eye view from uncalibrated street-level imagery with deep learning [2.7195102129095003]
We introduce a simple approach for estimating a bird's eye view from images without prior knowledge of a given camera's intrinsic and extrinsic parameters.
The framework has been applied to several applications to generate a live Map from camera feeds and to analyse social distancing violations at the city scale.
arXiv Detail & Related papers (2024-12-18T21:55:58Z) - Low-Light Image Enhancement Framework for Improved Object Detection in Fisheye Lens Datasets [4.170227455727819]
This study addresses the evolving challenges in urban traffic monitoring systems based on fisheye lens cameras.
Fisheye lenses provide wide and omnidirectional coverage in a single frame, making them a transformative solution.
Motivated by these challenges, this study proposes a novel approach that combines a ransformer-based image enhancement framework and ensemble learning technique.
arXiv Detail & Related papers (2024-04-15T18:32:52Z) - A Holistic Framework Towards Vision-based Traffic Signal Control with
Microscopic Simulation [53.39174966020085]
Traffic signal control (TSC) is crucial for reducing traffic congestion that leads to smoother traffic flow, reduced idling time, and mitigated CO2 emissions.
In this study, we explore the computer vision approach for TSC that modulates on-road traffic flows through visual observation.
We introduce a holistic traffic simulation framework called TrafficDojo towards vision-based TSC and its benchmarking.
arXiv Detail & Related papers (2024-03-11T16:42:29Z) - Application of 2D Homography for High Resolution Traffic Data Collection
using CCTV Cameras [9.946460710450319]
This study implements a three-stage video analytics framework for extracting high-resolution traffic data from CCTV cameras.
The key components of the framework include object recognition, perspective transformation, and vehicle trajectory reconstruction.
The results of the study showed about +/- 4.5% error rate for directional traffic counts, less than 10% MSE for speed bias between camera estimates.
arXiv Detail & Related papers (2024-01-14T07:33:14Z) - Camera-Radar Perception for Autonomous Vehicles and ADAS: Concepts,
Datasets and Metrics [77.34726150561087]
This work aims to carry out a study on the current scenario of camera and radar-based perception for ADAS and autonomous vehicles.
Concepts and characteristics related to both sensors, as well as to their fusion, are presented.
We give an overview of the Deep Learning-based detection and segmentation tasks, and the main datasets, metrics, challenges, and open questions in vehicle perception.
arXiv Detail & Related papers (2023-03-08T00:48:32Z) - Street-View Image Generation from a Bird's-Eye View Layout [95.36869800896335]
Bird's-Eye View (BEV) Perception has received increasing attention in recent years.
Data-driven simulation for autonomous driving has been a focal point of recent research.
We propose BEVGen, a conditional generative model that synthesizes realistic and spatially consistent surrounding images.
arXiv Detail & Related papers (2023-01-11T18:39:34Z) - Multi-Camera Multi-Object Tracking on the Move via Single-Stage Global
Association Approach [23.960847268459293]
This work introduces novel Single-Stage Global Association Tracking approaches to associate one or more detection from multi-cameras with tracked objects.
Our models also improve the detection accuracy of the standard vision-based 3D object detectors in the nuScenes detection challenge.
arXiv Detail & Related papers (2022-11-17T17:03:24Z) - Monocular BEV Perception of Road Scenes via Front-to-Top View Projection [57.19891435386843]
We present a novel framework that reconstructs a local map formed by road layout and vehicle occupancy in the bird's-eye view.
Our model runs at 25 FPS on a single GPU, which is efficient and applicable for real-time panorama HD map reconstruction.
arXiv Detail & Related papers (2022-11-15T13:52:41Z) - Multi-Camera Multiple 3D Object Tracking on the Move for Autonomous
Vehicles [17.12321292167318]
It is important for object detection and tracking to address new challenges, such as achieving consistent results across views of cameras.
This work presents a new Global Association Graph Model with Link Prediction approach to predict existing tracklets location and link detections with tracklets.
Our model exploits to improve the detection accuracy of a standard 3D object detector in the nuScenes detection challenge.
arXiv Detail & Related papers (2022-04-19T22:50:36Z) - Scalable and Real-time Multi-Camera Vehicle Detection,
Re-Identification, and Tracking [58.95210121654722]
We propose a real-time city-scale multi-camera vehicle tracking system that handles real-world, low-resolution CCTV instead of idealized and curated video streams.
Our method is ranked among the top five performers on the public leaderboard.
arXiv Detail & Related papers (2022-04-15T12:47:01Z) - Multi-Modal Fusion Transformer for End-to-End Autonomous Driving [59.60483620730437]
We propose TransFuser, a novel Multi-Modal Fusion Transformer, to integrate image and LiDAR representations using attention.
Our approach achieves state-of-the-art driving performance while reducing collisions by 76% compared to geometry-based fusion.
arXiv Detail & Related papers (2021-04-19T11:48:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.