Lidar-Monocular Surface Reconstruction Using Line Segments
- URL: http://arxiv.org/abs/2104.02761v1
- Date: Tue, 6 Apr 2021 19:49:53 GMT
- Title: Lidar-Monocular Surface Reconstruction Using Line Segments
- Authors: Victor Amblard, Timothy P. Osedach, Arnaud Croux, Andrew Speck and
John J. Leonard
- Abstract summary: We propose to leverage common geometric features that are detected in both the LIDAR scans and image data, allowing data from the two sensors to be processed in a higher-level space.
We show that our method delivers results that are comparable to a state-of-the-art LIDAR survey while not requiring highly accurate ground truth pose estimates.
- Score: 5.542669744873386
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Structure from Motion (SfM) often fails to estimate accurate poses in
environments that lack suitable visual features. In such cases, the quality of
the final 3D mesh, which is contingent on the accuracy of those estimates, is
reduced. One way to overcome this problem is to combine data from a monocular
camera with that of a LIDAR. This allows fine details and texture to be
captured while still accurately representing featureless subjects. However,
fusing these two sensor modalities is challenging due to their fundamentally
different characteristics. Rather than directly fusing image features and LIDAR
points, we propose to leverage common geometric features that are detected in
both the LIDAR scans and image data, allowing data from the two sensors to be
processed in a higher-level space. In particular, we propose to find
correspondences between 3D lines extracted from LIDAR scans and 2D lines
detected in images before performing a bundle adjustment to refine poses. We
also exploit the detected and optimized line segments to improve the quality of
the final mesh. We test our approach on the recently published dataset, Newer
College Dataset. We compare the accuracy and the completeness of the 3D mesh to
a ground truth obtained with a survey-grade 3D scanner. We show that our method
delivers results that are comparable to a state-of-the-art LIDAR survey while
not requiring highly accurate ground truth pose estimates.
Related papers
- PF3plat: Pose-Free Feed-Forward 3D Gaussian Splatting [54.7468067660037]
PF3plat sets a new state-of-the-art across all benchmarks, supported by comprehensive ablation studies validating our design choices.
Our framework capitalizes on fast speed, scalability, and high-quality 3D reconstruction and view synthesis capabilities of 3DGS.
arXiv Detail & Related papers (2024-10-29T15:28:15Z) - Approaching Outside: Scaling Unsupervised 3D Object Detection from 2D Scene [22.297964850282177]
We propose LiDAR-2D Self-paced Learning (LiSe) for unsupervised 3D detection.
RGB images serve as a valuable complement to LiDAR data, offering precise 2D localization cues.
Our framework devises a self-paced learning pipeline that incorporates adaptive sampling and weak model aggregation strategies.
arXiv Detail & Related papers (2024-07-11T14:58:49Z) - SplatPose & Detect: Pose-Agnostic 3D Anomaly Detection [18.796625355398252]
State-of-the-art algorithms are able to detect defects in increasingly difficult settings and data modalities.
We propose the novel 3D Gaussian splatting-based framework SplatPose which accurately estimates the pose of unseen views in a differentiable manner.
We achieve state-of-the-art results in both training and inference speed, and detection performance, even when using less training data than competing methods.
arXiv Detail & Related papers (2024-04-10T08:48:09Z) - 3DiffTection: 3D Object Detection with Geometry-Aware Diffusion Features [70.50665869806188]
3DiffTection is a state-of-the-art method for 3D object detection from single images.
We fine-tune a diffusion model to perform novel view synthesis conditioned on a single image.
We further train the model on target data with detection supervision.
arXiv Detail & Related papers (2023-11-07T23:46:41Z) - Semantics-aware LiDAR-Only Pseudo Point Cloud Generation for 3D Object
Detection [0.7234862895932991]
Recent advances introduced pseudo-LiDAR, i.e., synthetic dense point clouds, using additional modalities such as cameras to enhance 3D object detection.
We present a novel LiDAR-only framework that augments raw scans with dense pseudo point clouds by relying on LiDAR sensors and scene semantics.
arXiv Detail & Related papers (2023-09-16T09:18:47Z) - LiDAR-CS Dataset: LiDAR Point Cloud Dataset with Cross-Sensors for 3D
Object Detection [36.77084564823707]
deep learning methods heavily rely on annotated data and often face domain generalization issues.
LiDAR-CS dataset is the first dataset that addresses the sensor-related gaps in the domain of 3D object detection in real traffic.
arXiv Detail & Related papers (2023-01-29T19:10:35Z) - Homography Loss for Monocular 3D Object Detection [54.04870007473932]
A differentiable loss function, termed as Homography Loss, is proposed to achieve the goal, which exploits both 2D and 3D information.
Our method yields the best performance compared with the other state-of-the-arts by a large margin on KITTI 3D datasets.
arXiv Detail & Related papers (2022-04-02T03:48:03Z) - MonoDistill: Learning Spatial Features for Monocular 3D Object Detection [80.74622486604886]
We propose a simple and effective scheme to introduce the spatial information from LiDAR signals to the monocular 3D detectors.
We use the resulting data to train a 3D detector with the same architecture as the baseline model.
Experimental results show that the proposed method can significantly boost the performance of the baseline model.
arXiv Detail & Related papers (2022-01-26T09:21:41Z) - Cylindrical and Asymmetrical 3D Convolution Networks for LiDAR-based
Perception [122.53774221136193]
State-of-the-art methods for driving-scene LiDAR-based perception often project the point clouds to 2D space and then process them via 2D convolution.
A natural remedy is to utilize the 3D voxelization and 3D convolution network.
We propose a new framework for the outdoor LiDAR segmentation, where cylindrical partition and asymmetrical 3D convolution networks are designed to explore the 3D geometric pattern.
arXiv Detail & Related papers (2021-09-12T06:25:11Z) - Uncertainty-Aware Camera Pose Estimation from Points and Lines [101.03675842534415]
Perspective-n-Point-and-Line (Pn$PL) aims at fast, accurate and robust camera localizations with respect to a 3D model from 2D-3D feature coordinates.
arXiv Detail & Related papers (2021-07-08T15:19:36Z) - Monocular Camera Localization in Prior LiDAR Maps with 2D-3D Line
Correspondences [16.34334330572825]
We propose an efficient monocular camera localization method in prior LiDAR maps using direct 2D-3D line correspondences.
With the pose prediction from VIO, we can efficiently obtain coarse 2D-3D line correspondences.
The proposed method can efficiently estimate camera poses without accumulated drifts or pose jumps in structured environments.
arXiv Detail & Related papers (2020-04-01T23:36:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.