Indoor Layout Estimation by 2D LiDAR and Camera Fusion
- URL: http://arxiv.org/abs/2001.05422v1
- Date: Wed, 15 Jan 2020 16:43:35 GMT
- Title: Indoor Layout Estimation by 2D LiDAR and Camera Fusion
- Authors: Jieyu Li, Robert L Stevenson
- Abstract summary: This paper presents an algorithm for indoor layout estimation and reconstruction through the fusion of a sequence of captured images and LiDAR data sets.
In the proposed system, a movable platform collects both intensity images and 2D LiDAR information.
- Score: 3.2387553628943535
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper presents an algorithm for indoor layout estimation and
reconstruction through the fusion of a sequence of captured images and LiDAR
data sets. In the proposed system, a movable platform collects both intensity
images and 2D LiDAR information. Pose estimation and semantic segmentation is
computed jointly by aligning the LiDAR points to line segments from the images.
For indoor scenes with walls orthogonal to floor, the alignment problem is
decoupled into top-down view projection and a 2D similarity transformation
estimation and solved by the recursive random sample consensus (R-RANSAC)
algorithm. Hypotheses can be generated, evaluated and optimized by integrating
new scans as the platform moves throughout the environment. The proposed method
avoids the need of extensive prior training or a cuboid layout assumption,
which is more effective and practical compared to most previous indoor layout
estimation methods. Multi-sensor fusion allows the capability of providing
accurate depth estimation and high resolution visual information.
Related papers
- CoCPF: Coordinate-based Continuous Projection Field for Ill-Posed Inverse Problem in Imaging [78.734927709231]
Sparse-view computed tomography (SVCT) reconstruction aims to acquire CT images based on sparsely-sampled measurements.
Due to ill-posedness, implicit neural representation (INR) techniques may leave considerable holes'' (i.e., unmodeled spaces) in their fields, leading to sub-optimal results.
We propose the Coordinate-based Continuous Projection Field (CoCPF), which aims to build hole-free representation fields for SVCT reconstruction.
arXiv Detail & Related papers (2024-06-21T08:38:30Z) - 360 Layout Estimation via Orthogonal Planes Disentanglement and
Multi-view Geometric Consistency Perception [60.23832277827669]
Existing panoramic layout estimation solutions tend to recover room boundaries from a vertically compressed sequence.
We present an unsupervised adaptation technique tailored for horizon-depth and ratio representations.
We also introduce an optimization strategy for decision-level layout analysis and a 1D cost volume construction method for feature-level multi-view aggregation.
arXiv Detail & Related papers (2023-12-26T12:16:03Z) - Handbook on Leveraging Lines for Two-View Relative Pose Estimation [82.72686460985297]
We propose an approach for estimating the relative pose between image pairs by jointly exploiting points, lines, and their coincidences in a hybrid manner.
Our hybrid framework combines the advantages of all configurations, enabling robust and accurate estimation in challenging environments.
arXiv Detail & Related papers (2023-09-27T21:43:04Z) - Multi-Scale Estimation for Omni-Directional Saliency Maps Using
Learnable Equator Bias [1.413861804135093]
Saliency maps represent probability distributions of gazing points with a head-mounted display.
This paper proposes a novel saliency-map estimation model for the omni-directional images.
The accuracy of the saliency maps was improved by the proposed method.
arXiv Detail & Related papers (2023-09-15T04:08:20Z) - Contour Context: Abstract Structural Distribution for 3D LiDAR Loop
Detection and Metric Pose Estimation [31.968749056155467]
This paper proposes a simple, effective, and efficient topological loop closure detection pipeline with accurate 3-DoF metric pose estimation.
We interpret the Cartesian birds' eye view (BEV) image projected from 3D LiDAR points as layered distribution of structures.
A retrieval key is designed to accelerate the search of a database indexed by layered KD-trees.
arXiv Detail & Related papers (2023-02-13T07:18:24Z) - Multi-Camera Collaborative Depth Prediction via Consistent Structure
Estimation [75.99435808648784]
We propose a novel multi-camera collaborative depth prediction method.
It does not require large overlapping areas while maintaining structure consistency between cameras.
Experimental results on DDAD and NuScenes datasets demonstrate the superior performance of our method.
arXiv Detail & Related papers (2022-10-05T03:44:34Z) - 2D LiDAR and Camera Fusion Using Motion Cues for Indoor Layout
Estimation [2.6905021039717987]
A ground robot explores an indoor space with a single floor and vertical walls, and collects a sequence of intensity images and 2D LiDAR datasets.
The alignment of sensor outputs and image segmentation are computed jointly by aligning LiDAR points.
The ambiguity in images for ground-wall boundary extraction is removed with the assistance of LiDAR observations.
arXiv Detail & Related papers (2022-04-24T06:26:02Z) - BinsFormer: Revisiting Adaptive Bins for Monocular Depth Estimation [46.678016537618845]
We present a novel framework called BinsFormer, tailored for the classification-regression-based depth estimation.
It mainly focuses on two crucial components in the specific task: 1) proper generation of adaptive bins and 2) sufficient interaction between probability distribution and bins predictions.
Experiments on the KITTI, NYU, and SUN RGB-D datasets demonstrate that BinsFormer surpasses state-of-the-art monocular depth estimation methods with prominent margins.
arXiv Detail & Related papers (2022-04-03T04:38:02Z) - Similarity-Aware Fusion Network for 3D Semantic Segmentation [87.51314162700315]
We propose a similarity-aware fusion network (SAFNet) to adaptively fuse 2D images and 3D point clouds for 3D semantic segmentation.
We employ a late fusion strategy where we first learn the geometric and contextual similarities between the input and back-projected (from 2D pixels) point clouds.
We show that SAFNet significantly outperforms existing state-of-the-art fusion-based approaches across various data integrity.
arXiv Detail & Related papers (2021-07-04T09:28:18Z) - LED2-Net: Monocular 360 Layout Estimation via Differentiable Depth
Rendering [59.63979143021241]
We formulate the task of 360 layout estimation as a problem of predicting depth on the horizon line of a panorama.
We propose the Differentiable Depth Rendering procedure to make the conversion from layout to depth prediction differentiable.
Our method achieves state-of-the-art performance on numerous 360 layout benchmark datasets.
arXiv Detail & Related papers (2021-04-01T15:48:41Z) - Pixel-Pair Occlusion Relationship Map(P2ORM): Formulation, Inference &
Application [20.63938300312815]
We formalize concepts around geometric occlusion in 2D images (i.e., ignoring semantics)
We propose a novel unified formulation of both occlusion boundaries and occlusion orientations via a pixel-pair occlusion relation.
Experiments on a variety of datasets demonstrate that our method outperforms existing ones on this task.
We also propose a new depth map refinement method that consistently improve the performance of state-of-the-art monocular depth estimation methods.
arXiv Detail & Related papers (2020-07-23T15:52:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.