Urban Traffic Surveillance (UTS): A fully probabilistic 3D tracking
approach based on 2D detections
- URL: http://arxiv.org/abs/2105.14993v2
- Date: Tue, 1 Jun 2021 12:59:51 GMT
- Title: Urban Traffic Surveillance (UTS): A fully probabilistic 3D tracking
approach based on 2D detections
- Authors: Henry Bradler, Adrian Kretz and Rudolf Mester
- Abstract summary: Urban Traffic Surveillance (UTS) is a surveillance system based on a monocular and calibrated video camera.
UTS tracks the vehicles using a 3D bounding box representation and a physically reasonable 3D motion model.
- Score: 11.34426502082293
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Urban Traffic Surveillance (UTS) is a surveillance system based on a
monocular and calibrated video camera that detects vehicles in an urban traffic
scenario with dense traffic on multiple lanes and vehicles performing sharp
turning maneuvers. UTS then tracks the vehicles using a 3D bounding box
representation and a physically reasonable 3D motion model relying on an
unscented Kalman filter based approach. Since UTS recovers positions, shape and
motion information in a three-dimensional world coordinate system, it can be
employed to recognize diverse traffic violations or to supply intelligent
vehicles with valuable traffic information. We build on YOLOv3 as a detector
yielding 2D bounding boxes and class labels for each vehicle. A 2D detector
renders our system much more independent to different camera perspectives as a
variety of labeled training data is available. This allows for a good
generalization while also being more hardware efficient. The task of 3D
tracking based on 2D detections is supported by integrating class specific
prior knowledge about the vehicle shape. We quantitatively evaluate UTS using
self generated synthetic data and ground truth from the CARLA simulator, due to
the non-existence of datasets with an urban vehicle surveillance setting and
labeled 3D bounding boxes. Additionally, we give a qualitative impression of
how UTS performs on real-world data. Our implementation is capable of operating
in real time on a reasonably modern workstation. To the best of our knowledge,
UTS is to date the only 3D vehicle tracking system in a surveillance scenario
(static camera observing moving targets).
Related papers
- HeightFormer: A Semantic Alignment Monocular 3D Object Detection Method from Roadside Perspective [11.841338298700421]
We propose a novel 3D object detection framework integrating Spatial Former and Voxel Pooling Former to enhance 2D-to-3D projection based on height estimation.
Experiments were conducted using the Rope3D and DAIR-V2X-I dataset, and the results demonstrated the outperformance of the proposed algorithm in the detection of both vehicles and cyclists.
arXiv Detail & Related papers (2024-10-10T09:37:33Z) - 3D Data Augmentation for Driving Scenes on Camera [50.41413053812315]
We propose a 3D data augmentation approach termed Drive-3DAug, aiming at augmenting the driving scenes on camera in the 3D space.
We first utilize Neural Radiance Field (NeRF) to reconstruct the 3D models of background and foreground objects.
Then, augmented driving scenes can be obtained by placing the 3D objects with adapted location and orientation at the pre-defined valid region of backgrounds.
arXiv Detail & Related papers (2023-03-18T05:51:05Z) - HUM3DIL: Semi-supervised Multi-modal 3D Human Pose Estimation for
Autonomous Driving [95.42203932627102]
3D human pose estimation is an emerging technology, which can enable the autonomous vehicle to perceive and understand the subtle and complex behaviors of pedestrians.
Our method efficiently makes use of these complementary signals, in a semi-supervised fashion and outperforms existing methods with a large margin.
Specifically, we embed LiDAR points into pixel-aligned multi-modal features, which we pass through a sequence of Transformer refinement stages.
arXiv Detail & Related papers (2022-12-15T11:15:14Z) - Real-Time And Robust 3D Object Detection with Roadside LiDARs [20.10416681832639]
We design a 3D object detection model that can detect traffic participants in roadside LiDARs in real-time.
Our model uses an existing 3D detector as a baseline and improves its accuracy.
We make a significant contribution with our LiDAR-based 3D detector that can be used for smart city applications.
arXiv Detail & Related papers (2022-07-11T21:33:42Z) - Weakly Supervised Training of Monocular 3D Object Detectors Using Wide
Baseline Multi-view Traffic Camera Data [19.63193201107591]
7DoF prediction of vehicles at an intersection is an important task for assessing potential conflicts between road users.
We develop an approach using a weakly supervised method of fine tuning 3D object detectors for traffic observation cameras.
Our method achieves vehicle 7DoF pose prediction accuracy on our dataset comparable to the top performing monocular 3D object detectors on autonomous vehicle datasets.
arXiv Detail & Related papers (2021-10-21T08:26:48Z) - Learnable Online Graph Representations for 3D Multi-Object Tracking [156.58876381318402]
We propose a unified and learning based approach to the 3D MOT problem.
We employ a Neural Message Passing network for data association that is fully trainable.
We show the merit of the proposed approach on the publicly available nuScenes dataset by achieving state-of-the-art performance of 65.6% AMOTA and 58% fewer ID-switches.
arXiv Detail & Related papers (2021-04-23T17:59:28Z) - Monocular Quasi-Dense 3D Object Tracking [99.51683944057191]
A reliable and accurate 3D tracking framework is essential for predicting future locations of surrounding objects and planning the observer's actions in numerous applications such as autonomous driving.
We propose a framework that can effectively associate moving objects over time and estimate their full 3D bounding box information from a sequence of 2D images captured on a moving platform.
arXiv Detail & Related papers (2021-03-12T15:30:02Z) - Probabilistic 3D Multi-Modal, Multi-Object Tracking for Autonomous
Driving [22.693895321632507]
We propose a probabilistic, multi-modal, multi-object tracking system consisting of different trainable modules.
We show that our method outperforms current state-of-the-art on the NuScenes Tracking dataset.
arXiv Detail & Related papers (2020-12-26T15:00:54Z) - Fine-Grained Vehicle Perception via 3D Part-Guided Visual Data
Augmentation [77.60050239225086]
We propose an effective training data generation process by fitting a 3D car model with dynamic parts to vehicles in real images.
Our approach is fully automatic without any human interaction.
We present a multi-task network for VUS parsing and a multi-stream network for VHI parsing.
arXiv Detail & Related papers (2020-12-15T03:03:38Z) - Training-free Monocular 3D Event Detection System for Traffic
Surveillance [93.65240041833319]
Existing event detection systems are mostly learning-based and have achieved convincing performance when a large amount of training data is available.
In real-world scenarios, collecting sufficient labeled training data is expensive and sometimes impossible.
We propose a training-free monocular 3D event detection system for traffic surveillance.
arXiv Detail & Related papers (2020-02-01T04:42:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.