Pedestrian Detection in 3D Point Clouds using Deep Neural Networks
- URL: http://arxiv.org/abs/2105.01151v1
- Date: Mon, 3 May 2021 20:12:11 GMT
- Title: Pedestrian Detection in 3D Point Clouds using Deep Neural Networks
- Authors: \`Oscar Lorente, Josep R. Casas, Santiago Royo, Ivan Caminal
- Abstract summary: We propose a PointNet++ architecture to detect pedestrians in dense 3D point clouds.
The aim is to explore the potential contribution of geometric information alone in pedestrian detection systems.
We also present a semi-automatic labeling system that transfers pedestrian and non-pedestrian labels from RGB images onto the 3D domain.
- Score: 2.6763498831034034
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Detecting pedestrians is a crucial task in autonomous driving systems to
ensure the safety of drivers and pedestrians. The technologies involved in
these algorithms must be precise and reliable, regardless of environment
conditions. Relying solely on RGB cameras may not be enough to recognize road
environments in situations where cameras cannot capture scenes properly. Some
approaches aim to compensate for these limitations by combining RGB cameras
with TOF sensors, such as LIDARs. However, there are few works that address
this problem using exclusively the 3D geometric information provided by LIDARs.
In this paper, we propose a PointNet++ based architecture to detect pedestrians
in dense 3D point clouds. The aim is to explore the potential contribution of
geometric information alone in pedestrian detection systems. We also present a
semi-automatic labeling system that transfers pedestrian and non-pedestrian
labels from RGB images onto the 3D domain. The fact that our datasets have RGB
registered with point clouds enables label transferring by back projection from
2D bounding boxes to point clouds, with only a light manual supervision to
validate results. We train PointNet++ with the geometry of the resulting 3D
labelled clusters. The evaluation confirms the effectiveness of the proposed
method, yielding precision and recall values around 98%.
Related papers
- Sparse Points to Dense Clouds: Enhancing 3D Detection with Limited LiDAR Data [68.18735997052265]
We propose a balanced approach that combines the advantages of monocular and point cloud-based 3D detection.
Our method requires only a small number of 3D points, that can be obtained from a low-cost, low-resolution sensor.
The accuracy of 3D detection improves by 20% compared to the state-of-the-art monocular detection methods.
arXiv Detail & Related papers (2024-04-10T03:54:53Z) - A9 Intersection Dataset: All You Need for Urban 3D Camera-LiDAR Roadside
Perception [20.10416681832639]
A9 Intersection dataset consists of labeled LiDAR point clouds and synchronized camera images.
Our dataset consists of 4.8k images and point clouds with more than 57.4k manually labeled 3D boxes.
arXiv Detail & Related papers (2023-06-15T16:39:51Z) - Multi-Modal 3D Object Detection by Box Matching [109.43430123791684]
We propose a novel Fusion network by Box Matching (FBMNet) for multi-modal 3D detection.
With the learned assignments between 3D and 2D object proposals, the fusion for detection can be effectively performed by combing their ROI features.
arXiv Detail & Related papers (2023-05-12T18:08:51Z) - Image-to-Lidar Self-Supervised Distillation for Autonomous Driving Data [80.14669385741202]
We propose a self-supervised pre-training method for 3D perception models tailored to autonomous driving data.
We leverage the availability of synchronized and calibrated image and Lidar sensors in autonomous driving setups.
Our method does not require any point cloud nor image annotations.
arXiv Detail & Related papers (2022-03-30T12:40:30Z) - Progressive Coordinate Transforms for Monocular 3D Object Detection [52.00071336733109]
We propose a novel and lightweight approach, dubbed em Progressive Coordinate Transforms (PCT) to facilitate learning coordinate representations.
In this paper, we propose a novel and lightweight approach, dubbed em Progressive Coordinate Transforms (PCT) to facilitate learning coordinate representations.
arXiv Detail & Related papers (2021-08-12T15:22:33Z) - FGR: Frustum-Aware Geometric Reasoning for Weakly Supervised 3D Vehicle
Detection [81.79171905308827]
We propose frustum-aware geometric reasoning (FGR) to detect vehicles in point clouds without any 3D annotations.
Our method consists of two stages: coarse 3D segmentation and 3D bounding box estimation.
It is able to accurately detect objects in 3D space with only 2D bounding boxes and sparse point clouds.
arXiv Detail & Related papers (2021-05-17T07:29:55Z) - Ground-aware Monocular 3D Object Detection for Autonomous Driving [6.5702792909006735]
Estimating the 3D position and orientation of objects in the environment with a single RGB camera is a challenging task for low-cost urban autonomous driving and mobile robots.
Most of the existing algorithms are based on the geometric constraints in 2D-3D correspondence, which stems from generic 6D object pose estimation.
We introduce a novel neural network module to fully utilize such application-specific priors in the framework of deep learning.
arXiv Detail & Related papers (2021-02-01T08:18:24Z) - RoIFusion: 3D Object Detection from LiDAR and Vision [7.878027048763662]
We propose a novel fusion algorithm by projecting a set of 3D Region of Interests (RoIs) from the point clouds to the 2D RoIs of the corresponding the images.
Our approach achieves state-of-the-art performance on the KITTI 3D object detection challenging benchmark.
arXiv Detail & Related papers (2020-09-09T20:23:27Z) - Single-Shot 3D Detection of Vehicles from Monocular RGB Images via
Geometry Constrained Keypoints in Real-Time [6.82446891805815]
We propose a novel 3D single-shot object detection method for detecting vehicles in monocular RGB images.
Our approach lifts 2D detections to 3D space by predicting additional regression and classification parameters.
We test our approach on different datasets for autonomous driving and evaluate it using the challenging KITTI 3D Object Detection and the novel nuScenes Object Detection benchmarks.
arXiv Detail & Related papers (2020-06-23T15:10:19Z) - ZoomNet: Part-Aware Adaptive Zooming Neural Network for 3D Object
Detection [69.68263074432224]
We present a novel framework named ZoomNet for stereo imagery-based 3D detection.
The pipeline of ZoomNet begins with an ordinary 2D object detection model which is used to obtain pairs of left-right bounding boxes.
To further exploit the abundant texture cues in RGB images for more accurate disparity estimation, we introduce a conceptually straight-forward module -- adaptive zooming.
arXiv Detail & Related papers (2020-03-01T17:18:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.