Generalizable Multi-Camera 3D Pedestrian Detection
- URL: http://arxiv.org/abs/2104.05813v1
- Date: Mon, 12 Apr 2021 20:58:25 GMT
- Title: Generalizable Multi-Camera 3D Pedestrian Detection
- Authors: Jo\~ao Paulo Lima, Rafael Roberto, Lucas Figueiredo, Francisco
Sim\~oes, Veronica Teichrieb
- Abstract summary: We present a multi-camera 3D pedestrian detection method that does not need to train using data from the target scene.
We estimate pedestrian location on the ground plane using a novel based on human body poses and person's bounding boxes from an off-the-shelf monocular detector.
We then project these locations onto the world ground plane and fuse them with a new formulation of a clique cover problem.
- Score: 1.8303072203996347
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a multi-camera 3D pedestrian detection method that does not need
to train using data from the target scene. We estimate pedestrian location on
the ground plane using a novel heuristic based on human body poses and person's
bounding boxes from an off-the-shelf monocular detector. We then project these
locations onto the world ground plane and fuse them with a new formulation of a
clique cover problem. We also propose an optional step for exploiting
pedestrian appearance during fusion by using a domain-generalizable person
re-identification model. We evaluated the proposed approach on the challenging
WILDTRACK dataset. It obtained a MODA of 0.569 and an F-score of 0.78, superior
to state-of-the-art generalizable detection techniques.
Related papers
- Sparse Points to Dense Clouds: Enhancing 3D Detection with Limited LiDAR Data [68.18735997052265]
We propose a balanced approach that combines the advantages of monocular and point cloud-based 3D detection.
Our method requires only a small number of 3D points, that can be obtained from a low-cost, low-resolution sensor.
The accuracy of 3D detection improves by 20% compared to the state-of-the-art monocular detection methods.
arXiv Detail & Related papers (2024-04-10T03:54:53Z) - Find n' Propagate: Open-Vocabulary 3D Object Detection in Urban Environments [67.83787474506073]
We tackle the limitations of current LiDAR-based 3D object detection systems.
We introduce a universal textscFind n' Propagate approach for 3D OV tasks.
We achieve up to a 3.97-fold increase in Average Precision (AP) for novel object classes.
arXiv Detail & Related papers (2024-03-20T12:51:30Z) - Generalized Few-Shot 3D Object Detection of LiDAR Point Cloud for
Autonomous Driving [91.39625612027386]
We propose a novel task, called generalized few-shot 3D object detection, where we have a large amount of training data for common (base) objects, but only a few data for rare (novel) classes.
Specifically, we analyze in-depth differences between images and point clouds, and then present a practical principle for the few-shot setting in the 3D LiDAR dataset.
To solve this task, we propose an incremental fine-tuning method to extend existing 3D detection models to recognize both common and rare objects.
arXiv Detail & Related papers (2023-02-08T07:11:36Z) - Ground Plane Matters: Picking Up Ground Plane Prior in Monocular 3D
Object Detection [92.75961303269548]
The ground plane prior is a very informative geometry clue in monocular 3D object detection (M3OD)
We propose a Ground Plane Enhanced Network (GPENet) which resolves both issues at one go.
Our GPENet can outperform other methods and achieve state-of-the-art performance, well demonstrating the effectiveness and the superiority of the proposed approach.
arXiv Detail & Related papers (2022-11-03T02:21:35Z) - Det6D: A Ground-Aware Full-Pose 3D Object Detector for Improving Terrain
Robustness [1.4620086904601473]
We propose Det6D, the first full-degree-of-freedom 3D object detector without spatial and postural limitations.
To predict full-degree poses, including pitch and roll, we design a ground-aware orientation branch.
Experiments on various datasets demonstrate the effectiveness and robustness of our method in different terrains.
arXiv Detail & Related papers (2022-07-19T17:12:48Z) - Multiview Detection with Cardboard Human Modeling [23.072791405965415]
We propose a new pedestrian representation scheme based on human point clouds modeling.
Specifically, using ray tracing for holistic human depth estimation, we model pedestrians as upright, thin cardboard point clouds on the ground.
arXiv Detail & Related papers (2022-07-05T12:47:26Z) - Progressive Coordinate Transforms for Monocular 3D Object Detection [52.00071336733109]
We propose a novel and lightweight approach, dubbed em Progressive Coordinate Transforms (PCT) to facilitate learning coordinate representations.
In this paper, we propose a novel and lightweight approach, dubbed em Progressive Coordinate Transforms (PCT) to facilitate learning coordinate representations.
arXiv Detail & Related papers (2021-08-12T15:22:33Z) - Gated3D: Monocular 3D Object Detection From Temporal Illumination Cues [28.806932489163888]
We propose a novel 3D object detection modality that exploits temporal illumination cues from a low-cost monocular gated imager.
We assess the proposed method on a novel 3D detection dataset that includes gated imagery captured in over 10,000 km of driving data.
arXiv Detail & Related papers (2021-02-06T16:06:51Z) - Single-Shot 3D Detection of Vehicles from Monocular RGB Images via
Geometry Constrained Keypoints in Real-Time [6.82446891805815]
We propose a novel 3D single-shot object detection method for detecting vehicles in monocular RGB images.
Our approach lifts 2D detections to 3D space by predicting additional regression and classification parameters.
We test our approach on different datasets for autonomous driving and evaluate it using the challenging KITTI 3D Object Detection and the novel nuScenes Object Detection benchmarks.
arXiv Detail & Related papers (2020-06-23T15:10:19Z) - DOPS: Learning to Detect 3D Objects and Predict their 3D Shapes [54.239416488865565]
We propose a fast single-stage 3D object detection method for LIDAR data.
The core novelty of our method is a fast, single-pass architecture that both detects objects in 3D and estimates their shapes.
We find that our proposed method achieves state-of-the-art results by 5% on object detection in ScanNet scenes, and it gets top results by 3.4% in the Open dataset.
arXiv Detail & Related papers (2020-04-02T17:48:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.