2D LiDAR and Camera Fusion Using Motion Cues for Indoor Layout
Estimation
- URL: http://arxiv.org/abs/2204.11202v1
- Date: Sun, 24 Apr 2022 06:26:02 GMT
- Title: 2D LiDAR and Camera Fusion Using Motion Cues for Indoor Layout
Estimation
- Authors: Jieyu Li, Robert Stevenson
- Abstract summary: A ground robot explores an indoor space with a single floor and vertical walls, and collects a sequence of intensity images and 2D LiDAR datasets.
The alignment of sensor outputs and image segmentation are computed jointly by aligning LiDAR points.
The ambiguity in images for ground-wall boundary extraction is removed with the assistance of LiDAR observations.
- Score: 2.6905021039717987
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: This paper presents a novel indoor layout estimation system based on the
fusion of 2D LiDAR and intensity camera data. A ground robot explores an indoor
space with a single floor and vertical walls, and collects a sequence of
intensity images and 2D LiDAR datasets. The LiDAR provides accurate depth
information, while the camera captures high-resolution data for semantic
interpretation. The alignment of sensor outputs and image segmentation are
computed jointly by aligning LiDAR points, as samples of the room contour, to
ground-wall boundaries in the images. The alignment problem is decoupled into a
top-down view projection and a 2D similarity transformation estimation, which
can be solved according to the vertical vanishing point and motion of two
sensors. The recursive random sample consensus algorithm is implemented to
generate, evaluate and optimize multiple hypotheses with the sequential
measurements. The system allows jointly analyzing the geometric interpretation
from different sensors without offline calibration. The ambiguity in images for
ground-wall boundary extraction is removed with the assistance of LiDAR
observations, which improves the accuracy of semantic segmentation. The
localization and mapping is refined using the fused data, which enables the
system to work reliably in scenes with low texture or low geometric features.
Related papers
- CVCP-Fusion: On Implicit Depth Estimation for 3D Bounding Box Prediction [2.0375637582248136]
Cross-View Center Point-Fusion is a state-of-the-art model to perform 3D object detection.
Our architecture utilizes aspects from previously established algorithms, Cross-View Transformers and CenterPoint.
arXiv Detail & Related papers (2024-10-15T02:55:07Z) - SCIPaD: Incorporating Spatial Clues into Unsupervised Pose-Depth Joint Learning [17.99904937160487]
We introduce SCIPaD, a novel approach that incorporates spatial clues for unsupervised depth-pose joint learning.
SCIPaD achieves a reduction of 22.2% in average translation error and 34.8% in average angular error for camera pose estimation task on the KITTI Odometry dataset.
arXiv Detail & Related papers (2024-07-07T06:52:51Z) - RGB-based Category-level Object Pose Estimation via Decoupled Metric
Scale Recovery [72.13154206106259]
We propose a novel pipeline that decouples the 6D pose and size estimation to mitigate the influence of imperfect scales on rigid transformations.
Specifically, we leverage a pre-trained monocular estimator to extract local geometric information.
A separate branch is designed to directly recover the metric scale of the object based on category-level statistics.
arXiv Detail & Related papers (2023-09-19T02:20:26Z) - Multi-Scale Estimation for Omni-Directional Saliency Maps Using
Learnable Equator Bias [1.413861804135093]
Saliency maps represent probability distributions of gazing points with a head-mounted display.
This paper proposes a novel saliency-map estimation model for the omni-directional images.
The accuracy of the saliency maps was improved by the proposed method.
arXiv Detail & Related papers (2023-09-15T04:08:20Z) - Monocular 3D Object Detection with Depth from Motion [74.29588921594853]
We take advantage of camera ego-motion for accurate object depth estimation and detection.
Our framework, named Depth from Motion (DfM), then uses the established geometry to lift 2D image features to the 3D space and detects 3D objects thereon.
Our framework outperforms state-of-the-art methods by a large margin on the KITTI benchmark.
arXiv Detail & Related papers (2022-07-26T15:48:46Z) - CorrI2P: Deep Image-to-Point Cloud Registration via Dense Correspondence [51.91791056908387]
We propose the first feature-based dense correspondence framework for addressing the image-to-point cloud registration problem, dubbed CorrI2P.
Specifically, given a pair of a 2D image before a 3D point cloud, we first transform them into high-dimensional feature space feed the features into a symmetric overlapping region to determine the region where the image point cloud overlap.
arXiv Detail & Related papers (2022-07-12T11:49:31Z) - Uncertainty-Aware Camera Pose Estimation from Points and Lines [101.03675842534415]
Perspective-n-Point-and-Line (Pn$PL) aims at fast, accurate and robust camera localizations with respect to a 3D model from 2D-3D feature coordinates.
arXiv Detail & Related papers (2021-07-08T15:19:36Z) - DeepI2P: Image-to-Point Cloud Registration via Deep Classification [71.3121124994105]
DeepI2P is a novel approach for cross-modality registration between an image and a point cloud.
Our method estimates the relative rigid transformation between the coordinate frames of the camera and Lidar.
We circumvent the difficulty by converting the registration problem into a classification and inverse camera projection optimization problem.
arXiv Detail & Related papers (2021-04-08T04:27:32Z) - Lidar-Monocular Surface Reconstruction Using Line Segments [5.542669744873386]
We propose to leverage common geometric features that are detected in both the LIDAR scans and image data, allowing data from the two sensors to be processed in a higher-level space.
We show that our method delivers results that are comparable to a state-of-the-art LIDAR survey while not requiring highly accurate ground truth pose estimates.
arXiv Detail & Related papers (2021-04-06T19:49:53Z) - PLUME: Efficient 3D Object Detection from Stereo Images [95.31278688164646]
Existing methods tackle the problem in two steps: first depth estimation is performed, a pseudo LiDAR point cloud representation is computed from the depth estimates, and then object detection is performed in 3D space.
We propose a model that unifies these two tasks in the same metric space.
Our approach achieves state-of-the-art performance on the challenging KITTI benchmark, with significantly reduced inference time compared with existing methods.
arXiv Detail & Related papers (2021-01-17T05:11:38Z) - Indoor Layout Estimation by 2D LiDAR and Camera Fusion [3.2387553628943535]
This paper presents an algorithm for indoor layout estimation and reconstruction through the fusion of a sequence of captured images and LiDAR data sets.
In the proposed system, a movable platform collects both intensity images and 2D LiDAR information.
arXiv Detail & Related papers (2020-01-15T16:43:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.