LiDAR-guided Stereo Matching with a Spatial Consistency Constraint
- URL: http://arxiv.org/abs/2202.09953v2
- Date: Thu, 24 Feb 2022 13:00:26 GMT
- Title: LiDAR-guided Stereo Matching with a Spatial Consistency Constraint
- Authors: Yongjun Zhang, Siyuan Zou, Xinyi Liu, Xu Huang, Yi Wan, and Yongxiang
Yao
- Abstract summary: This study proposes an innovative LiDAR-guided stereo matching approach called LiDAR-guided stereo matching (LGSM)
The LGSM first detects the homogeneous pixels of each LiDAR projection point based on their color or intensity similarity.
Our formulation expands the constraint of sparse LiDAR projection points with the guidance of image information to optimize the cost volume of pixels.
- Score: 7.448893867145506
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The complementary fusion of light detection and ranging (LiDAR) data and
image data is a promising but challenging task for generating high-precision
and high-density point clouds. This study proposes an innovative LiDAR-guided
stereo matching approach called LiDAR-guided stereo matching (LGSM), which
considers the spatial consistency represented by continuous disparity or depth
changes in the homogeneous region of an image. The LGSM first detects the
homogeneous pixels of each LiDAR projection point based on their color or
intensity similarity. Next, we propose a riverbed enhancement function to
optimize the cost volume of the LiDAR projection points and their homogeneous
pixels to improve the matching robustness. Our formulation expands the
constraint scopes of sparse LiDAR projection points with the guidance of image
information to optimize the cost volume of pixels as much as possible. We
applied LGSM to semi-global matching and AD-Census on both simulated and real
datasets. When the percentage of LiDAR points in the simulated datasets was
0.16%, the matching accuracy of our method achieved a subpixel level, while
that of the original stereo matching algorithm was 3.4 pixels. The experimental
results show that LGSM is suitable for indoor, street, aerial, and satellite
image datasets and provides good transferability across semi-global matching
and AD-Census. Furthermore, the qualitative and quantitative evaluations
demonstrate that LGSM is superior to two state-of-the-art optimizing cost
volume methods, especially in reducing mismatches in difficult matching areas
and refining the boundaries of objects.
Related papers
- RSGaussian:3D Gaussian Splatting with LiDAR for Aerial Remote Sensing Novel View Synthesis [6.900071309404811]
RSGaussian is an innovative novel view synthesis (NVS) method for aerial remote sensing scenes.
It incorporates LiDAR point cloud as constraints into the 3D Gaussian Splatting method, which ensures that Gaussians grow and split along geometric benchmarks.
The approach also introduces coordinate transformations with distortion parameters for camera models to achieve pixel-level alignment between LiDAR point clouds and 2D images.
arXiv Detail & Related papers (2024-12-24T12:08:50Z) - LiDAR-GS:Real-time LiDAR Re-Simulation using Gaussian Splatting [50.808933338389686]
LiDAR simulation plays a crucial role in closed-loop simulation for autonomous driving.
We present LiDAR-GS, the first LiDAR Gaussian Splatting method, for real-time high-fidelity re-simulation of LiDAR sensor scans in public urban road scenes.
Our approach succeeds in simultaneously re-simulating depth, intensity, and ray-drop channels, achieving state-of-the-art results in both rendering frame rate and quality on publically available large scene datasets.
arXiv Detail & Related papers (2024-10-07T15:07:56Z) - Gaussian Splatting with Localized Points Management [52.009874685460694]
Localized Point Management (LPM) is capable of identifying those error-contributing zones in the highest demand for both point addition and geometry calibration.
LPM applies point densification in the identified zone, whilst resetting the opacity of those points residing in front of these regions so that a new opportunity is created to correct ill-conditioned points.
Notably, LPM improves both vanilla 3DGS and SpaceTimeGS to achieve state-of-the-art rendering quality while retaining real-time speeds.
arXiv Detail & Related papers (2024-06-06T16:55:07Z) - LIP-Loc: LiDAR Image Pretraining for Cross-Modal Localization [0.9562145896371785]
We apply Contrastive Language-Image Pre-Training to the domains of 2D image and 3D LiDAR points on the task of cross-modal localization.
Our method outperforms state-of-the-art recall@1 accuracy on the KITTI-360 dataset by 22.4%, using only perspective images.
We also demonstrate the zero-shot capabilities of our model and we beat SOTA by 8% without even training on it.
arXiv Detail & Related papers (2023-12-27T17:23:57Z) - LiDAR-Camera Panoptic Segmentation via Geometry-Consistent and
Semantic-Aware Alignment [63.83894701779067]
We propose LCPS, the first LiDAR-Camera Panoptic network.
In our approach, we conduct LiDAR-Camera fusion in three stages.
Our fusion strategy improves about 6.9% PQ performance over the LiDAR-only baseline on NuScenes dataset.
arXiv Detail & Related papers (2023-08-03T10:57:58Z) - From One to Many: Dynamic Cross Attention Networks for LiDAR and Camera
Fusion [12.792769704561024]
Existing fusion methods tend to align each 3D point to only one projected image pixel based on calibration.
We propose a Dynamic Cross Attention (DCA) module with a novel one-to-many cross-modality mapping.
The whole fusion architecture named Dynamic Cross Attention Network (DCAN) exploits multi-level image features and adapts to multiple representations of point clouds.
arXiv Detail & Related papers (2022-09-25T16:10:14Z) - Boosting 3D Object Detection by Simulating Multimodality on Point Clouds [51.87740119160152]
This paper presents a new approach to boost a single-modality (LiDAR) 3D object detector by teaching it to simulate features and responses that follow a multi-modality (LiDAR-image) detector.
The approach needs LiDAR-image data only when training the single-modality detector, and once well-trained, it only needs LiDAR data at inference.
Experimental results on the nuScenes dataset show that our approach outperforms all SOTA LiDAR-only 3D detectors.
arXiv Detail & Related papers (2022-06-30T01:44:30Z) - Calibrating LiDAR and Camera using Semantic Mutual information [8.40460868324361]
We propose an algorithm for automatic, targetless, extrinsic calibration of a LiDAR and camera system using semantic information.
We achieve this goal by maximizing mutual information (MI) of semantic information between sensors, leveraging a neural network to estimate semantic mutual information, and matrix exponential for calibration computation.
arXiv Detail & Related papers (2021-04-24T21:04:33Z) - Leveraging Spatial and Photometric Context for Calibrated Non-Lambertian
Photometric Stereo [61.6260594326246]
We introduce an efficient fully-convolutional architecture that can leverage both spatial and photometric context simultaneously.
Using separable 4D convolutions and 2D heat-maps reduces the size and makes more efficient.
arXiv Detail & Related papers (2021-03-22T18:06:58Z) - Fast and Robust Registration of Aerial Images and LiDAR data Based on
Structrual Features and 3D Phase Correlation [6.3812295314207335]
This paper proposes an automatic registration method based on structural features and three-dimension (3D) phase correlation.
Experiments with two datasets of aerial images and LiDAR data show that the proposed method is much faster and more robust than state of the art methods.
arXiv Detail & Related papers (2020-04-21T08:19:56Z) - Multi-View Photometric Stereo: A Robust Solution and Benchmark Dataset
for Spatially Varying Isotropic Materials [65.95928593628128]
We present a method to capture both 3D shape and spatially varying reflectance with a multi-view photometric stereo technique.
Our algorithm is suitable for perspective cameras and nearby point light sources.
arXiv Detail & Related papers (2020-01-18T12:26:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.