A9 Intersection Dataset: All You Need for Urban 3D Camera-LiDAR Roadside
Perception
- URL: http://arxiv.org/abs/2306.09266v1
- Date: Thu, 15 Jun 2023 16:39:51 GMT
- Title: A9 Intersection Dataset: All You Need for Urban 3D Camera-LiDAR Roadside
Perception
- Authors: Walter Zimmer, Christian Cre{\ss}, Huu Tung Nguyen, Alois C. Knoll
- Abstract summary: A9 Intersection dataset consists of labeled LiDAR point clouds and synchronized camera images.
Our dataset consists of 4.8k images and point clouds with more than 57.4k manually labeled 3D boxes.
- Score: 20.10416681832639
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Intelligent Transportation Systems (ITS) allow a drastic expansion of the
visibility range and decrease occlusions for autonomous driving. To obtain
accurate detections, detailed labeled sensor data for training is required.
Unfortunately, high-quality 3D labels of LiDAR point clouds from the
infrastructure perspective of an intersection are still rare. Therefore, we
provide the A9 Intersection Dataset, which consists of labeled LiDAR point
clouds and synchronized camera images. Here, we recorded the sensor output from
two roadside cameras and LiDARs mounted on intersection gantry bridges. The
point clouds were labeled in 3D by experienced annotators. Furthermore, we
provide calibration data between all sensors, which allow the projection of the
3D labels into the camera images and an accurate data fusion. Our dataset
consists of 4.8k images and point clouds with more than 57.4k manually labeled
3D boxes. With ten object classes, it has a high diversity of road users in
complex driving maneuvers, such as left and right turns, overtaking, and
U-turns. In experiments, we provided multiple baselines for the perception
tasks. Overall, our dataset is a valuable contribution to the scientific
community to perform complex 3D camera-LiDAR roadside perception tasks. Find
data, code, and more information at https://a9-dataset.com.
Related papers
- Sparse Points to Dense Clouds: Enhancing 3D Detection with Limited LiDAR Data [68.18735997052265]
We propose a balanced approach that combines the advantages of monocular and point cloud-based 3D detection.
Our method requires only a small number of 3D points, that can be obtained from a low-cost, low-resolution sensor.
The accuracy of 3D detection improves by 20% compared to the state-of-the-art monocular detection methods.
arXiv Detail & Related papers (2024-04-10T03:54:53Z) - TUMTraf V2X Cooperative Perception Dataset [20.907021313266128]
We propose CoopDet3D, a cooperative multi-modal fusion model, and TUMTraf-V2X, a perception dataset.
Our dataset contains 2,000 labeled point clouds and 5,000 labeled images from five roadside and four onboard sensors.
We show that our CoopDet3D camera-LiDAR fusion model achieves an increase of +14.36 3D mAP compared to a vehicle camera-LiDAR fusion model.
arXiv Detail & Related papers (2024-03-02T21:29:04Z) - InfraDet3D: Multi-Modal 3D Object Detection based on Roadside
Infrastructure Camera and LiDAR Sensors [23.058209168505247]
We introduce InfraDet3D, a multi-modal 3D object detector for roadside infrastructure sensors.
We fuse two LiDARs using early fusion and further incorporate detections from monocular cameras to increase robustness and to detect small objects.
The perception framework is deployed on a real-world intersection that is part of the A9 Test Stretch in Munich, Germany.
arXiv Detail & Related papers (2023-04-29T17:59:55Z) - 3D Data Augmentation for Driving Scenes on Camera [50.41413053812315]
We propose a 3D data augmentation approach termed Drive-3DAug, aiming at augmenting the driving scenes on camera in the 3D space.
We first utilize Neural Radiance Field (NeRF) to reconstruct the 3D models of background and foreground objects.
Then, augmented driving scenes can be obtained by placing the 3D objects with adapted location and orientation at the pre-defined valid region of backgrounds.
arXiv Detail & Related papers (2023-03-18T05:51:05Z) - SUPS: A Simulated Underground Parking Scenario Dataset for Autonomous
Driving [41.221988979184665]
SUPS is a simulated dataset for underground automatic parking.
It supports multiple tasks with multiple sensors and multiple semantic labels aligned with successive images.
We also evaluate the state-of-the-art SLAM algorithms and perception models on our dataset.
arXiv Detail & Related papers (2023-02-25T02:59:12Z) - Benchmarking the Robustness of LiDAR-Camera Fusion for 3D Object
Detection [58.81316192862618]
Two critical sensors for 3D perception in autonomous driving are the camera and the LiDAR.
fusing these two modalities can significantly boost the performance of 3D perception models.
We benchmark the state-of-the-art fusion methods for the first time.
arXiv Detail & Related papers (2022-05-30T09:35:37Z) - Image-to-Lidar Self-Supervised Distillation for Autonomous Driving Data [80.14669385741202]
We propose a self-supervised pre-training method for 3D perception models tailored to autonomous driving data.
We leverage the availability of synchronized and calibrated image and Lidar sensors in autonomous driving setups.
Our method does not require any point cloud nor image annotations.
arXiv Detail & Related papers (2022-03-30T12:40:30Z) - Rope3D: TheRoadside Perception Dataset for Autonomous Driving and
Monocular 3D Object Detection Task [48.555440807415664]
We present the first high-diversity challenging Roadside Perception 3D dataset- Rope3D from a novel view.
The dataset consists of 50k images and over 1.5M 3D objects in various scenes.
We propose to leverage the geometry constraint to solve the inherent ambiguities caused by various sensors, viewpoints.
arXiv Detail & Related papers (2022-03-25T12:13:23Z) - PC-DAN: Point Cloud based Deep Affinity Network for 3D Multi-Object
Tracking (Accepted as an extended abstract in JRDB-ACT Workshop at CVPR21) [68.12101204123422]
A point cloud is a dense compilation of spatial data in 3D coordinates.
We propose a PointNet-based approach for 3D Multi-Object Tracking (MOT)
arXiv Detail & Related papers (2021-06-03T05:36:39Z) - Pedestrian Detection in 3D Point Clouds using Deep Neural Networks [2.6763498831034034]
We propose a PointNet++ architecture to detect pedestrians in dense 3D point clouds.
The aim is to explore the potential contribution of geometric information alone in pedestrian detection systems.
We also present a semi-automatic labeling system that transfers pedestrian and non-pedestrian labels from RGB images onto the 3D domain.
arXiv Detail & Related papers (2021-05-03T20:12:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.