RegFormer: An Efficient Projection-Aware Transformer Network for
Large-Scale Point Cloud Registration
- URL: http://arxiv.org/abs/2303.12384v3
- Date: Thu, 10 Aug 2023 02:39:22 GMT
- Title: RegFormer: An Efficient Projection-Aware Transformer Network for
Large-Scale Point Cloud Registration
- Authors: Jiuming Liu, Guangming Wang, Zhe Liu, Chaokang Jiang, Marc Pollefeys,
Hesheng Wang
- Abstract summary: We propose an end-to-end transformer network (RegFormer) for large-scale point cloud alignment.
Specifically, a projection-aware hierarchical transformer is proposed to capture long-range dependencies and filter outliers.
Our transformer has linear complexity, which guarantees high efficiency even for large-scale scenes.
- Score: 73.69415797389195
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Although point cloud registration has achieved remarkable advances in
object-level and indoor scenes, large-scale registration methods are rarely
explored. Challenges mainly arise from the huge point number, complex
distribution, and outliers of outdoor LiDAR scans. In addition, most existing
registration works generally adopt a two-stage paradigm: They first find
correspondences by extracting discriminative local features and then leverage
estimators (eg. RANSAC) to filter outliers, which are highly dependent on
well-designed descriptors and post-processing choices. To address these
problems, we propose an end-to-end transformer network (RegFormer) for
large-scale point cloud alignment without any further post-processing.
Specifically, a projection-aware hierarchical transformer is proposed to
capture long-range dependencies and filter outliers by extracting point
features globally. Our transformer has linear complexity, which guarantees high
efficiency even for large-scale scenes. Furthermore, to effectively reduce
mismatches, a bijective association transformer is designed for regressing the
initial transformation. Extensive experiments on KITTI and NuScenes datasets
demonstrate that our RegFormer achieves competitive performance in terms of
both accuracy and efficiency.
Related papers
- MEDPNet: Achieving High-Precision Adaptive Registration for Complex Die Castings [10.504847830252254]
This paper proposes a high-precision adaptive registration method called Multiscale Efficient Deep Closest Point (MEDPNet)
The MEDPNet method performs coarse die-casting point cloud data registration using the Efficient-DCP method, followed by precision registration using the Multiscale feature fusion dual-channel registration (MDR) method.
Our proposed method demonstrates excellent performance compared to state-of-the-art geometric and learning-based registration methods when applied to complex die-casting point cloud data.
arXiv Detail & Related papers (2024-03-15T03:42:38Z) - Adaptive Point Transformer [88.28498667506165]
Adaptive Point Cloud Transformer (AdaPT) is a standard PT model augmented by an adaptive token selection mechanism.
AdaPT dynamically reduces the number of tokens during inference, enabling efficient processing of large point clouds.
arXiv Detail & Related papers (2024-01-26T13:24:45Z) - Pointerformer: Deep Reinforced Multi-Pointer Transformer for the
Traveling Salesman Problem [67.32731657297377]
Traveling Salesman Problem (TSP) is a classic routing optimization problem originally arising in the domain of transportation and logistics.
Recently, Deep Reinforcement Learning has been increasingly employed to solve TSP due to its high inference efficiency.
We propose a novel end-to-end DRL approach, referred to as Pointerformer, based on multi-pointer Transformer.
arXiv Detail & Related papers (2023-04-19T03:48:32Z) - Transformers for Object Detection in Large Point Clouds [9.287964414592826]
We present TransLPC, a novel detection model for large point clouds based on a transformer architecture.
We propose a novel query refinement technique to improve detection accuracy, while retaining a memory-friendly number of transformer decoder queries.
This simple technique has a significant effect on detection accuracy, which is evaluated on the challenging nuScenes dataset on real-world lidar data.
arXiv Detail & Related papers (2022-09-30T06:35:43Z) - CloudAttention: Efficient Multi-Scale Attention Scheme For 3D Point
Cloud Learning [81.85951026033787]
We set transformers in this work and incorporate them into a hierarchical framework for shape classification and part and scene segmentation.
We also compute efficient and dynamic global cross attentions by leveraging sampling and grouping at each iteration.
The proposed hierarchical model achieves state-of-the-art shape classification in mean accuracy and yields results on par with the previous segmentation methods.
arXiv Detail & Related papers (2022-07-31T21:39:15Z) - Full Transformer Framework for Robust Point Cloud Registration with Deep
Information Interaction [9.431484068349903]
Recent Transformer-based methods have achieved advanced performance in point cloud registration.
Recent CNNs fail to model global relations due to their local fields receptive.
shallow-wide architecture of Transformers and lack of positional encoding lead to indistinct feature extraction.
arXiv Detail & Related papers (2021-12-17T08:40:52Z) - Multimodality Biomedical Image Registration using Free Point Transformer
Networks [0.37501702548174964]
We describe a point-set registration algorithm based on a novel free point transformer (FPT) network.
FPT is constructed with a global feature extractor which accepts unordered source and target point-sets of variable size.
In a multimodal registration task using prostate MR and sparsely acquired ultrasound images, FPT yields comparable or improved results.
arXiv Detail & Related papers (2020-08-05T00:13:04Z) - The Cascade Transformer: an Application for Efficient Answer Sentence
Selection [116.09532365093659]
We introduce the Cascade Transformer, a technique to adapt transformer-based models into a cascade of rankers.
When compared to a state-of-the-art transformer model, our approach reduces computation by 37% with almost no impact on accuracy.
arXiv Detail & Related papers (2020-05-05T23:32:01Z) - Resolution Adaptive Networks for Efficient Inference [53.04907454606711]
We propose a novel Resolution Adaptive Network (RANet), which is inspired by the intuition that low-resolution representations are sufficient for classifying "easy" inputs.
In RANet, the input images are first routed to a lightweight sub-network that efficiently extracts low-resolution representations.
High-resolution paths in the network maintain the capability to recognize the "hard" samples.
arXiv Detail & Related papers (2020-03-16T16:54:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.