End-to-End Learning Local Multi-view Descriptors for 3D Point Clouds
- URL: http://arxiv.org/abs/2003.05855v2
- Date: Mon, 16 Mar 2020 14:32:05 GMT
- Title: End-to-End Learning Local Multi-view Descriptors for 3D Point Clouds
- Authors: Lei Li, Siyu Zhu, Hongbo Fu, Ping Tan, Chiew-Lan Tai
- Abstract summary: We propose an end-to-end framework to learn local multi-view descriptors for 3D point clouds.
Our method outperforms existing local descriptors both quantitatively and qualitatively.
- Score: 67.27510166559563
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this work, we propose an end-to-end framework to learn local multi-view
descriptors for 3D point clouds. To adopt a similar multi-view representation,
existing studies use hand-crafted viewpoints for rendering in a preprocessing
stage, which is detached from the subsequent descriptor learning stage. In our
framework, we integrate the multi-view rendering into neural networks by using
a differentiable renderer, which allows the viewpoints to be optimizable
parameters for capturing more informative local context of interest points. To
obtain discriminative descriptors, we also design a soft-view pooling module to
attentively fuse convolutional features across views. Extensive experiments on
existing 3D registration benchmarks show that our method outperforms existing
local descriptors both quantitatively and qualitatively.
Related papers
- MVTN: Learning Multi-View Transformations for 3D Understanding [60.15214023270087]
We introduce the Multi-View Transformation Network (MVTN), which uses differentiable rendering to determine optimal view-points for 3D shape recognition.
MVTN can be trained end-to-end with any multi-view network for 3D shape recognition.
Our approach demonstrates state-of-the-art performance in 3D classification and shape retrieval on several benchmarks.
arXiv Detail & Related papers (2022-12-27T12:09:16Z) - Voint Cloud: Multi-View Point Cloud Representation for 3D Understanding [80.04281842702294]
We introduce the concept of the multi-view point cloud (Voint cloud) representing each 3D point as a set of features extracted from several view-points.
This novel 3D Voint cloud representation combines the compactness of 3D point cloud representation with the natural view-awareness of multi-view representation.
We deploy a Voint neural network (VointNet) with a theoretically established functional form to learn representations in the Voint space.
arXiv Detail & Related papers (2021-11-30T13:08:19Z) - PnP-3D: A Plug-and-Play for 3D Point Clouds [38.05362492645094]
We propose a plug-and-play module, -3D, to improve the effectiveness of existing networks in analyzing point cloud data.
To thoroughly evaluate our approach, we conduct experiments on three standard point cloud analysis tasks.
In addition to achieving state-of-the-art results, we present comprehensive studies to demonstrate our approach's advantages.
arXiv Detail & Related papers (2021-08-16T23:59:43Z) - UPDesc: Unsupervised Point Descriptor Learning for Robust Registration [54.95201961399334]
UPDesc is an unsupervised method to learn point descriptors for robust point cloud registration.
We show that our learned descriptors yield superior performance over existing unsupervised methods.
arXiv Detail & Related papers (2021-08-05T17:11:08Z) - DH3D: Deep Hierarchical 3D Descriptors for Robust Large-Scale 6DoF
Relocalization [56.15308829924527]
We propose a Siamese network that jointly learns 3D local feature detection and description directly from raw 3D points.
For detecting 3D keypoints we predict the discriminativeness of the local descriptors in an unsupervised manner.
Experiments on various benchmarks demonstrate that our method achieves competitive results for both global point cloud retrieval and local point cloud registration.
arXiv Detail & Related papers (2020-07-17T20:21:22Z) - Learning and Matching Multi-View Descriptors for Registration of Point
Clouds [48.25586496457587]
We first propose a multi-view local descriptor, which is learned from the images of multiple views, for the description of 3D keypoints.
Then, we develop a robust matching approach, aiming at rejecting outlier matches based on the efficient inference.
We have demonstrated the boost of our approaches to registration on the public scanning and multi-view stereo datasets.
arXiv Detail & Related papers (2018-07-16T01:58:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.