A Flexible Multi-view Multi-modal Imaging System for Outdoor Scenes
- URL: http://arxiv.org/abs/2302.10465v1
- Date: Tue, 21 Feb 2023 06:14:05 GMT
- Title: A Flexible Multi-view Multi-modal Imaging System for Outdoor Scenes
- Authors: Meng Zhang, Wenxuan Guo, Bohao Fan, Yifan Chen, Jianjiang Feng and Jie
Zhou
- Abstract summary: We propose a wireless multi-view multi-modal 3D imaging system generally applicable to large outdoor scenes.
Multiple spatially distributed slave nodes equipped with cameras and LiDARs are connected to form a wireless sensor network.
This system is the first imaging system that integrates mutli-view RGB cameras and LiDARs in large outdoor scenes.
- Score: 37.5716419191546
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multi-view imaging systems enable uniform coverage of 3D space and reduce the
impact of occlusion, which is beneficial for 3D object detection and tracking
accuracy. However, existing imaging systems built with multi-view cameras or
depth sensors are limited by the small applicable scene and complicated
composition. In this paper, we propose a wireless multi-view multi-modal 3D
imaging system generally applicable to large outdoor scenes, which consists of
a master node and several slave nodes. Multiple spatially distributed slave
nodes equipped with cameras and LiDARs are connected to form a wireless sensor
network. While providing flexibility and scalability, the system applies
automatic spatio-temporal calibration techniques to obtain accurate 3D
multi-view multi-modal data. This system is the first imaging system that
integrates mutli-view RGB cameras and LiDARs in large outdoor scenes among
existing 3D imaging systems. We perform point clouds based 3D object detection
and long-term tracking using the 3D imaging dataset collected by this system.
The experimental results show that multi-view point clouds greatly improve 3D
object detection and tracking accuracy regardless of complex and various
outdoor environments.
Related papers
- VFMM3D: Releasing the Potential of Image by Vision Foundation Model for Monocular 3D Object Detection [80.62052650370416]
monocular 3D object detection holds significant importance across various applications, including autonomous driving and robotics.
In this paper, we present VFMM3D, an innovative framework that leverages the capabilities of Vision Foundation Models (VFMs) to accurately transform single-view images into LiDAR point cloud representations.
arXiv Detail & Related papers (2024-04-15T03:12:12Z) - Sparse Points to Dense Clouds: Enhancing 3D Detection with Limited LiDAR Data [68.18735997052265]
We propose a balanced approach that combines the advantages of monocular and point cloud-based 3D detection.
Our method requires only a small number of 3D points, that can be obtained from a low-cost, low-resolution sensor.
The accuracy of 3D detection improves by 20% compared to the state-of-the-art monocular detection methods.
arXiv Detail & Related papers (2024-04-10T03:54:53Z) - Multi-Modal Dataset Acquisition for Photometrically Challenging Object [56.30027922063559]
This paper addresses the limitations of current datasets for 3D vision tasks in terms of accuracy, size, realism, and suitable imaging modalities for photometrically challenging objects.
We propose a novel annotation and acquisition pipeline that enhances existing 3D perception and 6D object pose datasets.
arXiv Detail & Related papers (2023-08-21T10:38:32Z) - Weakly Supervised 3D Multi-person Pose Estimation for Large-scale Scenes
based on Monocular Camera and Single LiDAR [41.39277657279448]
We propose a monocular camera and single LiDAR-based method for 3D multi-person pose estimation in large-scale scenes.
Specifically, we design an effective fusion strategy to take advantage of multi-modal input data, including images and point cloud.
Our method exploits the inherent geometry constraints of point cloud for self-supervision and utilizes 2D keypoints on images for weak supervision.
arXiv Detail & Related papers (2022-11-30T12:50:40Z) - Scatter Points in Space: 3D Detection from Multi-view Monocular Images [8.71944437852952]
3D object detection from monocular image(s) is a challenging and long-standing problem of computer vision.
Recent methods tend to aggregate multiview feature by sampling regular 3D grid densely in space.
We propose a learnable keypoints sampling method, which scatters pseudo surface points in 3D space, in order to keep data sparsity.
arXiv Detail & Related papers (2022-08-31T09:38:05Z) - PC-DAN: Point Cloud based Deep Affinity Network for 3D Multi-Object
Tracking (Accepted as an extended abstract in JRDB-ACT Workshop at CVPR21) [68.12101204123422]
A point cloud is a dense compilation of spatial data in 3D coordinates.
We propose a PointNet-based approach for 3D Multi-Object Tracking (MOT)
arXiv Detail & Related papers (2021-06-03T05:36:39Z) - RoIFusion: 3D Object Detection from LiDAR and Vision [7.878027048763662]
We propose a novel fusion algorithm by projecting a set of 3D Region of Interests (RoIs) from the point clouds to the 2D RoIs of the corresponding the images.
Our approach achieves state-of-the-art performance on the KITTI 3D object detection challenging benchmark.
arXiv Detail & Related papers (2020-09-09T20:23:27Z) - Lightweight Multi-View 3D Pose Estimation through Camera-Disentangled
Representation [57.11299763566534]
We present a solution to recover 3D pose from multi-view images captured with spatially calibrated cameras.
We exploit 3D geometry to fuse input images into a unified latent representation of pose, which is disentangled from camera view-points.
Our architecture then conditions the learned representation on camera projection operators to produce accurate per-view 2d detections.
arXiv Detail & Related papers (2020-04-05T12:52:29Z) - A Bayesian Filter for Multi-view 3D Multi-object Tracking with Occlusion
Handling [2.824395407508717]
The proposed algorithm has a linear complexity in the total number of detections across the cameras.
It operates in the 3D world frame, and provides 3D trajectory estimates of the objects.
The proposed algorithm is evaluated on the latest WILDTRACKS dataset, and demonstrated to work in very crowded scenes.
arXiv Detail & Related papers (2020-01-13T09:34:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.