ParaNet: Deep Regular Representation for 3D Point Clouds
- URL: http://arxiv.org/abs/2012.03028v1
- Date: Sat, 5 Dec 2020 13:19:55 GMT
- Title: ParaNet: Deep Regular Representation for 3D Point Clouds
- Authors: Qijian Zhang, Junhui Hou, Yue Qian, Juyong Zhang, Ying He
- Abstract summary: ParaNet is a novel end-to-end deep learning framework for representing 3D point clouds.
It converts an irregular 3D point cloud into a regular 2D color image, named point geometry image (PGI)
In contrast to conventional regular representation modalities based on multi-view projection and voxelization, the proposed representation is differentiable and reversible.
- Score: 62.81379889095186
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Although convolutional neural networks have achieved remarkable success in
analyzing 2D images/videos, it is still non-trivial to apply the well-developed
2D techniques in regular domains to the irregular 3D point cloud data. To
bridge this gap, we propose ParaNet, a novel end-to-end deep learning
framework, for representing 3D point clouds in a completely regular and nearly
lossless manner. To be specific, ParaNet converts an irregular 3D point cloud
into a regular 2D color image, named point geometry image (PGI), where each
pixel encodes the spatial coordinates of a point. In contrast to conventional
regular representation modalities based on multi-view projection and
voxelization, the proposed representation is differentiable and reversible.
Technically, ParaNet is composed of a surface embedding module, which
parameterizes 3D surface points onto a unit square, and a grid resampling
module, which resamples the embedded 2D manifold over regular dense grids. Note
that ParaNet is unsupervised, i.e., the training simply relies on
reference-free geometry constraints. The PGIs can be seamlessly coupled with a
task network established upon standard and mature techniques for 2D
images/videos to realize a specific task for 3D point clouds. We evaluate
ParaNet over shape classification and point cloud upsampling, in which our
solutions perform favorably against the existing state-of-the-art methods. We
believe such a paradigm will open up many possibilities to advance the progress
of deep learning-based point cloud processing and understanding.
Related papers
- Flattening-Net: Deep Regular 2D Representation for 3D Point Cloud
Analysis [66.49788145564004]
We present an unsupervised deep neural architecture called Flattening-Net to represent irregular 3D point clouds of arbitrary geometry and topology.
Our methods perform favorably against the current state-of-the-art competitors.
arXiv Detail & Related papers (2022-12-17T15:05:25Z) - PointMCD: Boosting Deep Point Cloud Encoders via Multi-view Cross-modal
Distillation for 3D Shape Recognition [55.38462937452363]
We propose a unified multi-view cross-modal distillation architecture, including a pretrained deep image encoder as the teacher and a deep point encoder as the student.
By pair-wise aligning multi-view visual and geometric descriptors, we can obtain more powerful deep point encoders without exhausting and complicated network modification.
arXiv Detail & Related papers (2022-07-07T07:23:20Z) - Learning Multi-View Aggregation In the Wild for Large-Scale 3D Semantic
Segmentation [3.5939555573102853]
Recent works on 3D semantic segmentation propose to exploit the synergy between images and point clouds by processing each modality with a dedicated network.
We propose an end-to-end trainable multi-view aggregation model leveraging the viewing conditions of 3D points to merge features from images taken at arbitrary positions.
Our method can combine standard 2D and 3D networks and outperforms both 3D models operating on colorized point clouds and hybrid 2D/3D networks.
arXiv Detail & Related papers (2022-04-15T17:10:48Z) - A Conditional Point Diffusion-Refinement Paradigm for 3D Point Cloud
Completion [69.32451612060214]
Real-scanned 3D point clouds are often incomplete, and it is important to recover complete point clouds for downstream applications.
Most existing point cloud completion methods use Chamfer Distance (CD) loss for training.
We propose a novel Point Diffusion-Refinement (PDR) paradigm for point cloud completion.
arXiv Detail & Related papers (2021-12-07T06:59:06Z) - Unsupervised Learning of Fine Structure Generation for 3D Point Clouds
by 2D Projection Matching [66.98712589559028]
We propose an unsupervised approach for 3D point cloud generation with fine structures.
Our method can recover fine 3D structures from 2D silhouette images at different resolutions.
arXiv Detail & Related papers (2021-08-08T22:15:31Z) - Learning Geometry-Disentangled Representation for Complementary
Understanding of 3D Object Point Cloud [50.56461318879761]
We propose Geometry-Disentangled Attention Network (GDANet) for 3D image processing.
GDANet disentangles point clouds into contour and flat part of 3D objects, respectively denoted by sharp and gentle variation components.
Experiments on 3D object classification and segmentation benchmarks demonstrate that GDANet achieves the state-of-the-arts with fewer parameters.
arXiv Detail & Related papers (2020-12-20T13:35:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.