iMatcher: Improve matching in point cloud registration via local-to-global geometric consistency learning
- URL: http://arxiv.org/abs/2509.08982v1
- Date: Wed, 10 Sep 2025 20:25:57 GMT
- Title: iMatcher: Improve matching in point cloud registration via local-to-global geometric consistency learning
- Authors: Karim Slimani, Catherine Achard, Brahim Tamadazte,
- Abstract summary: iMatcher is a framework for feature matching in point cloud registration.<n>It uses both local and global consistency to predict a point-wise matching probability.<n>It achieves state-of-the-art inlier ratios, scoring 95% - 97% on KITTI, 94% - 97% on KITTI-360, and up to 81.1% on 3DMatch.
- Score: 2.3985192761907643
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper presents iMatcher, a fully differentiable framework for feature matching in point cloud registration. The proposed method leverages learned features to predict a geometrically consistent confidence matrix, incorporating both local and global consistency. First, a local graph embedding module leads to an initialization of the score matrix. A subsequent repositioning step refines this matrix by considering bilateral source-to-target and target-to-source matching via nearest neighbor search in 3D space. The paired point features are then stacked together to be refined through global geometric consistency learning to predict a point-wise matching probability. Extensive experiments on real-world outdoor (KITTI, KITTI-360) and indoor (3DMatch) datasets, as well as on 6-DoF pose estimation (TUD-L) and partial-to-partial matching (MVP-RG), demonstrate that iMatcher significantly improves rigid registration performance. The method achieves state-of-the-art inlier ratios, scoring 95% - 97% on KITTI, 94% - 97% on KITTI-360, and up to 81.1% on 3DMatch, highlighting its robustness across diverse settings.
Related papers
- Multiway Point Cloud Mosaicking with Diffusion and Global Optimization [74.3802812773891]
We introduce a novel framework for multiway point cloud mosaicking (named Wednesday)
At the core of our approach is ODIN, a learned pairwise registration algorithm that identifies overlaps and refines attention scores.
Tested on four diverse, large-scale datasets, our method state-of-the-art pairwise and rotation registration results by a large margin on all benchmarks.
arXiv Detail & Related papers (2024-03-30T17:29:13Z) - LFM-3D: Learnable Feature Matching Across Wide Baselines Using 3D
Signals [9.201550006194994]
Learnable matchers often underperform when there exists only small regions of co-visibility between image pairs.
We propose LFM-3D, a Learnable Feature Matching framework that uses models based on graph neural networks.
We show that the resulting improved correspondences lead to much higher relative posing accuracy for in-the-wild image pairs.
arXiv Detail & Related papers (2023-03-22T17:46:27Z) - Learning to Register Unbalanced Point Pairs [10.369750912567714]
Recent 3D registration methods can effectively handle large-scale or partially overlapping point pairs.
We present a novel 3D registration method, called UPPNet, for the unbalanced point pairs.
arXiv Detail & Related papers (2022-07-09T08:03:59Z) - BoxGraph: Semantic Place Recognition and Pose Estimation from 3D LiDAR [22.553026961366005]
We model 3D point clouds as fully-connected graphs of semantically identified components.
Optimal association across graphs allows for full 6-Degree-of-Freedom (DoF) pose estimation and place recognition.
This representation is very concise, condensing the size of maps by a factor of 25 against the state-of-the-art.
arXiv Detail & Related papers (2022-06-30T09:39:08Z) - REGTR: End-to-end Point Cloud Correspondences with Transformers [79.52112840465558]
We conjecture that attention mechanisms can replace the role of explicit feature matching and RANSAC.
We propose an end-to-end framework to directly predict the final set of correspondences.
Our approach achieves state-of-the-art performance on 3DMatch and ModelNet benchmarks.
arXiv Detail & Related papers (2022-03-28T06:01:00Z) - DFC: Deep Feature Consistency for Robust Point Cloud Registration [0.4724825031148411]
We present a novel learning-based alignment network for complex alignment scenes.
We validate our approach on the 3DMatch dataset and the KITTI odometry dataset.
arXiv Detail & Related papers (2021-11-15T08:27:21Z) - Deep Hough Voting for Robust Global Registration [52.40611370293272]
We present an efficient framework for pairwise registration of real-world 3D scans, leveraging Hough voting in the 6D transformation parameter space.
Our method outperforms state-of-the-art methods on 3DMatch and 3DLoMatch benchmarks while achieving comparable performance on KITTI odometry dataset.
arXiv Detail & Related papers (2021-09-09T14:38:06Z) - 3D Correspondence Grouping with Compatibility Features [51.869670613445685]
We present a simple yet effective method for 3D correspondence grouping.
The objective is to accurately classify initial correspondences obtained by matching local geometric descriptors into inliers and outliers.
We propose a novel representation for 3D correspondences, dubbed compatibility feature (CF), to describe the consistencies within inliers and inconsistencies within outliers.
arXiv Detail & Related papers (2020-07-21T02:39:48Z) - Deep Global Registration [90.05565444450524]
Deep Global Registration is a differentiable framework for pairwise registration of real-world 3D scans.
Our approach outperforms state-of-the-art methods, both learning-based and classical, on real-world data.
arXiv Detail & Related papers (2020-04-24T05:47:32Z) - Learning multiview 3D point cloud registration [74.39499501822682]
We present a novel, end-to-end learnable, multiview 3D point cloud registration algorithm.
Our approach outperforms the state-of-the-art by a significant margin, while being end-to-end trainable and computationally less costly.
arXiv Detail & Related papers (2020-01-15T03:42:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.