Implicit Autoencoder for Point-Cloud Self-Supervised Representation
Learning
- URL: http://arxiv.org/abs/2201.00785v5
- Date: Sun, 27 Aug 2023 18:01:10 GMT
- Title: Implicit Autoencoder for Point-Cloud Self-Supervised Representation
Learning
- Authors: Siming Yan, Zhenpei Yang, Haoxiang Li, Chen Song, Li Guan, Hao Kang,
Gang Hua, Qixing Huang
- Abstract summary: The most popular and accessible 3D representation, i.e., point clouds, involves discrete samples of the underlying continuous 3D surface.
This discretization process introduces sampling variations on the 3D shape, making it challenging to develop transferable knowledge of the true 3D geometry.
In the standard autoencoding paradigm, the encoder is compelled to encode not only the 3D geometry but also information on the specific discrete sampling of the 3D shape into the latent code.
This is because the point cloud reconstructed by the decoder is considered unacceptable unless there is a perfect mapping between the original and the reconstructed
- Score: 39.521374237630766
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper advocates the use of implicit surface representation in
autoencoder-based self-supervised 3D representation learning. The most popular
and accessible 3D representation, i.e., point clouds, involves discrete samples
of the underlying continuous 3D surface. This discretization process introduces
sampling variations on the 3D shape, making it challenging to develop
transferable knowledge of the true 3D geometry. In the standard autoencoding
paradigm, the encoder is compelled to encode not only the 3D geometry but also
information on the specific discrete sampling of the 3D shape into the latent
code. This is because the point cloud reconstructed by the decoder is
considered unacceptable unless there is a perfect mapping between the original
and the reconstructed point clouds. This paper introduces the Implicit
AutoEncoder (IAE), a simple yet effective method that addresses the sampling
variation issue by replacing the commonly-used point-cloud decoder with an
implicit decoder. The implicit decoder reconstructs a continuous representation
of the 3D shape, independent of the imperfections in the discrete samples.
Extensive experiments demonstrate that the proposed IAE achieves
state-of-the-art performance across various self-supervised learning
benchmarks.
Related papers
- DatasetNeRF: Efficient 3D-aware Data Factory with Generative Radiance Fields [68.94868475824575]
This paper introduces a novel approach capable of generating infinite, high-quality 3D-consistent 2D annotations alongside 3D point cloud segmentations.
We leverage the strong semantic prior within a 3D generative model to train a semantic decoder.
Once trained, the decoder efficiently generalizes across the latent space, enabling the generation of infinite data.
arXiv Detail & Related papers (2023-11-18T21:58:28Z) - 3D Feature Prediction for Masked-AutoEncoder-Based Point Cloud Pretraining [45.58631796379208]
Masked autoencoders (MAEs) have recently been introduced to 3D self-supervised pretraining for point clouds.
We propose to ignore point position reconstruction and recover high-order features at masked points through a novel attention-based decoder.
We validate our pretext task and decoder design using different encoder structures for 3D training and demonstrate the advantages of our pretrained networks on various point cloud analysis tasks.
arXiv Detail & Related papers (2023-04-14T03:25:24Z) - PointMCD: Boosting Deep Point Cloud Encoders via Multi-view Cross-modal
Distillation for 3D Shape Recognition [55.38462937452363]
We propose a unified multi-view cross-modal distillation architecture, including a pretrained deep image encoder as the teacher and a deep point encoder as the student.
By pair-wise aligning multi-view visual and geometric descriptors, we can obtain more powerful deep point encoders without exhausting and complicated network modification.
arXiv Detail & Related papers (2022-07-07T07:23:20Z) - SNAKE: Shape-aware Neural 3D Keypoint Field [62.91169625183118]
Detecting 3D keypoints from point clouds is important for shape reconstruction.
This work investigates the dual question: can shape reconstruction benefit 3D keypoint detection?
We propose a novel unsupervised paradigm named SNAKE, which is short for shape-aware neural 3D keypoint field.
arXiv Detail & Related papers (2022-06-03T17:58:43Z) - Self-Supervised Point Cloud Representation Learning with Occlusion
Auto-Encoder [63.77257588569852]
We present 3D Occlusion Auto-Encoder (3D-OAE) for learning representations for point clouds.
Our key idea is to randomly occlude some local patches of the input point cloud and establish the supervision via recovering the occluded patches.
In contrast with previous methods, our 3D-OAE can remove a large proportion of patches and predict them only with a small number of visible patches.
arXiv Detail & Related papers (2022-03-26T14:06:29Z) - Learning 3D Dense Correspondence via Canonical Point Autoencoder [108.20735652143787]
We propose a canonical point autoencoder (CPAE) that predicts dense correspondences between 3D shapes of the same category.
The autoencoder performs two key functions: (a) encoding an arbitrarily ordered point cloud to a canonical primitive, and (b) decoding the primitive back to the original input instance shape.
arXiv Detail & Related papers (2021-07-10T15:54:48Z) - PvDeConv: Point-Voxel Deconvolution for Autoencoding CAD Construction in
3D [23.87757211847093]
We learn to synthesize high-resolution point clouds of 10k points that densely describe the underlying geometry of Computer Aided Design (CAD) models.
We introduce a new dedicated dataset, the CC3D, containing 50k+ pairs of CAD models and their corresponding 3D meshes.
This dataset is used to learn a convolutional autoencoder for point clouds sampled from the pairs of 3D scans - CAD models.
arXiv Detail & Related papers (2021-01-12T14:14:13Z) - Exploring Deep 3D Spatial Encodings for Large-Scale 3D Scene
Understanding [19.134536179555102]
We propose an alternative approach to overcome the limitations of CNN based approaches by encoding the spatial features of raw 3D point clouds into undirected graph models.
The proposed method achieves on par state-of-the-art accuracy with improved training time and model stability thus indicating strong potential for further research.
arXiv Detail & Related papers (2020-11-29T12:56:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.