PanoNet3D: Combining Semantic and Geometric Understanding for LiDARPoint
Cloud Detection
- URL: http://arxiv.org/abs/2012.09418v1
- Date: Thu, 17 Dec 2020 06:58:34 GMT
- Title: PanoNet3D: Combining Semantic and Geometric Understanding for LiDARPoint
Cloud Detection
- Authors: Xia Chen, Jianren Wang, David Held, Martial Hebert
- Abstract summary: We propose to learn both semantic feature and geometric structure via a unified multi-view framework.
By fusing semantic and geometric features, our method outperforms state-of-the-art approaches in all categories by a large margin.
- Score: 40.907188672454986
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Visual data in autonomous driving perception, such as camera image and LiDAR
point cloud, can be interpreted as a mixture of two aspects: semantic feature
and geometric structure. Semantics come from the appearance and context of
objects to the sensor, while geometric structure is the actual 3D shape of
point clouds. Most detectors on LiDAR point clouds focus only on analyzing the
geometric structure of objects in real 3D space. Unlike previous works, we
propose to learn both semantic feature and geometric structure via a unified
multi-view framework. Our method exploits the nature of LiDAR scans -- 2D range
images, and applies well-studied 2D convolutions to extract semantic features.
By fusing semantic and geometric features, our method outperforms
state-of-the-art approaches in all categories by a large margin. The
methodology of combining semantic and geometric features provides a unique
perspective of looking at the problems in real-world 3D point cloud detection.
Related papers
- Geometry-guided Feature Learning and Fusion for Indoor Scene Reconstruction [14.225228781008209]
This paper proposes a novel geometry integration mechanism for 3D scene reconstruction.
Our approach incorporates 3D geometry at three levels, i.e. feature learning, feature fusion, and network supervision.
arXiv Detail & Related papers (2024-08-28T08:02:47Z) - Robust 3D Tracking with Quality-Aware Shape Completion [67.9748164949519]
We propose a synthetic target representation composed of dense and complete point clouds depicting the target shape precisely by shape completion for robust 3D tracking.
Specifically, we design a voxelized 3D tracking framework with shape completion, in which we propose a quality-aware shape completion mechanism to alleviate the adverse effect of noisy historical predictions.
arXiv Detail & Related papers (2023-12-17T04:50:24Z) - MS23D: A 3D Object Detection Method Using Multi-Scale Semantic Feature Points to Construct 3D Feature Layer [4.644319899528183]
LiDAR point clouds can effectively depict the motion and posture of objects in three-dimensional space.
In autonomous driving scenarios, the sparsity and hollowness of point clouds create some difficulties for voxel-based methods.
We propose a two-stage 3D object detection framework, called MS23D.
arXiv Detail & Related papers (2023-08-31T08:03:25Z) - PointMCD: Boosting Deep Point Cloud Encoders via Multi-view Cross-modal
Distillation for 3D Shape Recognition [55.38462937452363]
We propose a unified multi-view cross-modal distillation architecture, including a pretrained deep image encoder as the teacher and a deep point encoder as the student.
By pair-wise aligning multi-view visual and geometric descriptors, we can obtain more powerful deep point encoders without exhausting and complicated network modification.
arXiv Detail & Related papers (2022-07-07T07:23:20Z) - Self-Supervised Image Representation Learning with Geometric Set
Consistency [50.12720780102395]
We propose a method for self-supervised image representation learning under the guidance of 3D geometric consistency.
Specifically, we introduce 3D geometric consistency into a contrastive learning framework to enforce the feature consistency within image views.
arXiv Detail & Related papers (2022-03-29T08:57:33Z) - Cylindrical and Asymmetrical 3D Convolution Networks for LiDAR-based
Perception [122.53774221136193]
State-of-the-art methods for driving-scene LiDAR-based perception often project the point clouds to 2D space and then process them via 2D convolution.
A natural remedy is to utilize the 3D voxelization and 3D convolution network.
We propose a new framework for the outdoor LiDAR segmentation, where cylindrical partition and asymmetrical 3D convolution networks are designed to explore the 3D geometric pattern.
arXiv Detail & Related papers (2021-09-12T06:25:11Z) - Learning Geometry-Disentangled Representation for Complementary
Understanding of 3D Object Point Cloud [50.56461318879761]
We propose Geometry-Disentangled Attention Network (GDANet) for 3D image processing.
GDANet disentangles point clouds into contour and flat part of 3D objects, respectively denoted by sharp and gentle variation components.
Experiments on 3D object classification and segmentation benchmarks demonstrate that GDANet achieves the state-of-the-arts with fewer parameters.
arXiv Detail & Related papers (2020-12-20T13:35:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.