3D Unsupervised Region-Aware Registration Transformer
- URL: http://arxiv.org/abs/2110.03544v3
- Date: Thu, 22 Feb 2024 07:41:44 GMT
- Title: 3D Unsupervised Region-Aware Registration Transformer
- Authors: Yu Hao, Yi Fang
- Abstract summary: Learning robust point cloud registration models with deep neural networks has emerged as a powerful paradigm.
We propose a new design of 3D region partition module that is able to divide the input shape to different regions with a self-supervised 3D shape reconstruction loss.
Our experiments show that our 3D-URRT achieves superior registration performance over various benchmark datasets.
- Score: 13.137287695912633
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper concerns the research problem of point cloud registration to find
the rigid transformation to optimally align the source point set with the
target one. Learning robust point cloud registration models with deep neural
networks has emerged as a powerful paradigm, offering promising performance in
predicting the global geometric transformation for a pair of point sets.
Existing methods first leverage an encoder to regress the global shape
descriptor, which is then decoded into a shape-conditioned transformation via
concatenation-based conditioning. However, different regions of a 3D shape vary
in their geometric structures which makes it more sense that we have a
region-conditioned transformation instead of the shape-conditioned one. In this
paper, we define our 3D registration function through the introduction of a new
design of 3D region partition module that is able to divide the input shape to
different regions with a self-supervised 3D shape reconstruction loss without
the need for ground truth labels. We further propose the 3D shape transformer
module to efficiently and effectively capture short- and long-range geometric
dependencies for regions on the 3D shape Consequently, the region-aware decoder
module is proposed to predict the transformations for different regions
respectively. The global geometric transformation from the source point set to
the target one is then formed by the weighted fusion of region-aware
transformation. Compared to the state-of-the-art approaches, our experiments
show that our 3D-URRT achieves superior registration performance over various
benchmark datasets (e.g. ModelNet40).
Related papers
- Boosting Cross-Domain Point Classification via Distilling Relational Priors from 2D Transformers [59.0181939916084]
Traditional 3D networks mainly focus on local geometric details and ignore the topological structure between local geometries.
We propose a novel Priors Distillation (RPD) method to extract priors from the well-trained transformers on massive images.
Experiments on the PointDA-10 and the Sim-to-Real datasets verify that the proposed method consistently achieves the state-of-the-art performance of UDA for point cloud classification.
arXiv Detail & Related papers (2024-07-26T06:29:09Z) - Learning SO(3)-Invariant Semantic Correspondence via Local Shape Transform [62.27337227010514]
We introduce a novel self-supervised Rotation-Invariant 3D correspondence learner with Local Shape Transform, dubbed RIST.
RIST learns to establish dense correspondences between shapes even under challenging intra-class variations and arbitrary orientations.
RIST demonstrates state-of-the-art performances on 3D part label transfer and semantic keypoint transfer given arbitrarily rotated point cloud pairs.
arXiv Detail & Related papers (2024-04-17T08:09:25Z) - NeuSDFusion: A Spatial-Aware Generative Model for 3D Shape Completion, Reconstruction, and Generation [52.772319840580074]
3D shape generation aims to produce innovative 3D content adhering to specific conditions and constraints.
Existing methods often decompose 3D shapes into a sequence of localized components, treating each element in isolation.
We introduce a novel spatial-aware 3D shape generation framework that leverages 2D plane representations for enhanced 3D shape modeling.
arXiv Detail & Related papers (2024-03-27T04:09:34Z) - Geometry-Contrastive Transformer for Generalized 3D Pose Transfer [95.56457218144983]
The intuition of this work is to perceive the geometric inconsistency between the given meshes with the powerful self-attention mechanism.
We propose a novel geometry-contrastive Transformer that has an efficient 3D structured perceiving ability to the global geometric inconsistencies.
We present a latent isometric regularization module together with a novel semi-synthesized dataset for the cross-dataset 3D pose transfer task.
arXiv Detail & Related papers (2021-12-14T13:14:24Z) - Learning to generate shape from global-local spectra [0.0]
We build our method on top of recent advances on the so called shape-from-spectrum paradigm.
We consider the spectrum as a natural and ready to use representation to encode variability of the shapes.
Our results confirm the improvement of the proposed approach in comparison to existing and alternative methods.
arXiv Detail & Related papers (2021-08-04T16:39:56Z) - Rotation-Invariant Local-to-Global Representation Learning for 3D Point
Cloud [42.86112554931754]
We propose a local-to-global representation learning algorithm for 3D point cloud data.
Our model takes advantage of multi-level abstraction based on graph convolutional neural networks.
The proposed algorithm presents the state-of-the-art performance on the rotation-augmented 3D object recognition and segmentation benchmarks.
arXiv Detail & Related papers (2020-10-07T10:30:20Z) - Learning Local Neighboring Structure for Robust 3D Shape Representation [143.15904669246697]
Representation learning for 3D meshes is important in many computer vision and graphics applications.
We propose a local structure-aware anisotropic convolutional operation (LSA-Conv)
Our model produces significant improvement in 3D shape reconstruction compared to state-of-the-art methods.
arXiv Detail & Related papers (2020-04-21T13:40:03Z) - A Rotation-Invariant Framework for Deep Point Cloud Analysis [132.91915346157018]
We introduce a new low-level purely rotation-invariant representation to replace common 3D Cartesian coordinates as the network inputs.
Also, we present a network architecture to embed these representations into features, encoding local relations between points and their neighbors, and the global shape structure.
We evaluate our method on multiple point cloud analysis tasks, including shape classification, part segmentation, and shape retrieval.
arXiv Detail & Related papers (2020-03-16T14:04:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.