Dynamic 3D Scene Analysis by Point Cloud Accumulation
- URL: http://arxiv.org/abs/2207.12394v1
- Date: Mon, 25 Jul 2022 17:57:46 GMT
- Title: Dynamic 3D Scene Analysis by Point Cloud Accumulation
- Authors: Shengyu Huang, Zan Gojcic, Jiahui Huang, Andreas Wieser, Konrad
Schindler
- Abstract summary: Multi-beam LiDAR sensors are used on autonomous vehicles and mobile robots.
Each frame covers the scene sparsely, due to limited angular scanning resolution and occlusion.
We propose a method that exploits inductive biases of outdoor street scenes, including their geometric layout and object-level rigidity.
- Score: 32.491921765128936
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Multi-beam LiDAR sensors, as used on autonomous vehicles and mobile robots,
acquire sequences of 3D range scans ("frames"). Each frame covers the scene
sparsely, due to limited angular scanning resolution and occlusion. The
sparsity restricts the performance of downstream processes like semantic
segmentation or surface reconstruction. Luckily, when the sensor moves, frames
are captured from a sequence of different viewpoints. This provides
complementary information and, when accumulated in a common scene coordinate
frame, yields a denser sampling and a more complete coverage of the underlying
3D scene. However, often the scanned scenes contain moving objects. Points on
those objects are not correctly aligned by just undoing the scanner's
ego-motion. In the present paper, we explore multi-frame point cloud
accumulation as a mid-level representation of 3D scan sequences, and develop a
method that exploits inductive biases of outdoor street scenes, including their
geometric layout and object-level rigidity. Compared to state-of-the-art scene
flow estimators, our proposed approach aims to align all 3D points in a common
reference frame correctly accumulating the points on the individual objects.
Our approach greatly reduces the alignment errors on several benchmark
datasets. Moreover, the accumulated point clouds benefit high-level tasks like
surface reconstruction.
Related papers
- U3DS$^3$: Unsupervised 3D Semantic Scene Segmentation [19.706172244951116]
This paper presents U3DS$3$, as a step towards completely unsupervised point cloud segmentation for any holistic 3D scenes.
The initial step of our proposed approach involves generating superpoints based on the geometric characteristics of each scene.
We then undergo a learning process through a spatial clustering-based methodology, followed by iterative training using pseudo-labels generated in accordance with the cluster centroids.
arXiv Detail & Related papers (2023-11-10T12:05:35Z) - Semantics-aware LiDAR-Only Pseudo Point Cloud Generation for 3D Object
Detection [0.7234862895932991]
Recent advances introduced pseudo-LiDAR, i.e., synthetic dense point clouds, using additional modalities such as cameras to enhance 3D object detection.
We present a novel LiDAR-only framework that augments raw scans with dense pseudo point clouds by relying on LiDAR sensors and scene semantics.
arXiv Detail & Related papers (2023-09-16T09:18:47Z) - V-DETR: DETR with Vertex Relative Position Encoding for 3D Object
Detection [73.37781484123536]
We introduce a highly performant 3D object detector for point clouds using the DETR framework.
To address the limitation, we introduce a novel 3D Relative Position (3DV-RPE) method.
We show exceptional results on the challenging ScanNetV2 benchmark.
arXiv Detail & Related papers (2023-08-08T17:14:14Z) - Quadric Representations for LiDAR Odometry, Mapping and Localization [93.24140840537912]
Current LiDAR odometry, mapping and localization methods leverage point-wise representations of 3D scenes.
We propose a novel method of describing scenes using quadric surfaces, which are far more compact representations of 3D objects.
Our method maintains low latency and memory utility while achieving competitive, and even superior, accuracy.
arXiv Detail & Related papers (2023-04-27T13:52:01Z) - Tracking by 3D Model Estimation of Unknown Objects in Videos [122.56499878291916]
We argue that this representation is limited and instead propose to guide and improve 2D tracking with an explicit object representation.
Our representation tackles a complex long-term dense correspondence problem between all 3D points on the object for all video frames.
The proposed optimization minimizes a novel loss function to estimate the best 3D shape, texture, and 6DoF pose.
arXiv Detail & Related papers (2023-04-13T11:32:36Z) - CMR3D: Contextualized Multi-Stage Refinement for 3D Object Detection [57.44434974289945]
We propose Contextualized Multi-Stage Refinement for 3D Object Detection (CMR3D) framework.
Our framework takes a 3D scene as input and strives to explicitly integrate useful contextual information of the scene.
In addition to 3D object detection, we investigate the effectiveness of our framework for the problem of 3D object counting.
arXiv Detail & Related papers (2022-09-13T05:26:09Z) - Scatter Points in Space: 3D Detection from Multi-view Monocular Images [8.71944437852952]
3D object detection from monocular image(s) is a challenging and long-standing problem of computer vision.
Recent methods tend to aggregate multiview feature by sampling regular 3D grid densely in space.
We propose a learnable keypoints sampling method, which scatters pseudo surface points in 3D space, in order to keep data sparsity.
arXiv Detail & Related papers (2022-08-31T09:38:05Z) - PolarMix: A General Data Augmentation Technique for LiDAR Point Clouds [100.03877236181546]
PolarMix is a point cloud augmentation technique that is simple and generic.
It can work as plug-and-play for various 3D deep architectures and also performs well for unsupervised domain adaptation.
arXiv Detail & Related papers (2022-07-30T13:52:19Z) - HyperDet3D: Learning a Scene-conditioned 3D Object Detector [154.84798451437032]
We propose HyperDet3D to explore scene-conditioned prior knowledge for 3D object detection.
Our HyperDet3D achieves state-of-the-art results on the 3D object detection benchmark of the ScanNet and SUN RGB-D datasets.
arXiv Detail & Related papers (2022-04-12T07:57:58Z) - RoIFusion: 3D Object Detection from LiDAR and Vision [7.878027048763662]
We propose a novel fusion algorithm by projecting a set of 3D Region of Interests (RoIs) from the point clouds to the 2D RoIs of the corresponding the images.
Our approach achieves state-of-the-art performance on the KITTI 3D object detection challenging benchmark.
arXiv Detail & Related papers (2020-09-09T20:23:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.