Indoor Layout Estimation by 2D LiDAR and Camera Fusion
- URL: http://arxiv.org/abs/2001.05422v1
- Date: Wed, 15 Jan 2020 16:43:35 GMT
- Title: Indoor Layout Estimation by 2D LiDAR and Camera Fusion
- Authors: Jieyu Li, Robert L Stevenson
- Abstract summary: This paper presents an algorithm for indoor layout estimation and reconstruction through the fusion of a sequence of captured images and LiDAR data sets.
In the proposed system, a movable platform collects both intensity images and 2D LiDAR information.
- Score: 3.2387553628943535
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper presents an algorithm for indoor layout estimation and
reconstruction through the fusion of a sequence of captured images and LiDAR
data sets. In the proposed system, a movable platform collects both intensity
images and 2D LiDAR information. Pose estimation and semantic segmentation is
computed jointly by aligning the LiDAR points to line segments from the images.
For indoor scenes with walls orthogonal to floor, the alignment problem is
decoupled into top-down view projection and a 2D similarity transformation
estimation and solved by the recursive random sample consensus (R-RANSAC)
algorithm. Hypotheses can be generated, evaluated and optimized by integrating
new scans as the platform moves throughout the environment. The proposed method
avoids the need of extensive prior training or a cuboid layout assumption,
which is more effective and practical compared to most previous indoor layout
estimation methods. Multi-sensor fusion allows the capability of providing
accurate depth estimation and high resolution visual information.
Related papers
- No Pose, No Problem: Surprisingly Simple 3D Gaussian Splats from Sparse Unposed Images [100.80376573969045]
NoPoSplat is a feed-forward model capable of reconstructing 3D scenes parameterized by 3D Gaussians from multi-view images.
Our model achieves real-time 3D Gaussian reconstruction during inference.
This work makes significant advances in pose-free generalizable 3D reconstruction and demonstrates its applicability to real-world scenarios.
arXiv Detail & Related papers (2024-10-31T17:58:22Z) - PF3plat: Pose-Free Feed-Forward 3D Gaussian Splatting [54.7468067660037]
PF3plat sets a new state-of-the-art across all benchmarks, supported by comprehensive ablation studies validating our design choices.
Our framework capitalizes on fast speed, scalability, and high-quality 3D reconstruction and view synthesis capabilities of 3DGS.
arXiv Detail & Related papers (2024-10-29T15:28:15Z) - Robust Two-View Geometry Estimation with Implicit Differentiation [2.048226951354646]
We present a novel two-view geometry estimation framework.
It is based on a differentiable robust loss function fitting.
We evaluate our approach on the camera pose estimation task in both outdoor and indoor scenarios.
arXiv Detail & Related papers (2024-10-23T15:51:33Z) - CVCP-Fusion: On Implicit Depth Estimation for 3D Bounding Box Prediction [2.0375637582248136]
Cross-View Center Point-Fusion is a state-of-the-art model to perform 3D object detection.
Our architecture utilizes aspects from previously established algorithms, Cross-View Transformers and CenterPoint.
arXiv Detail & Related papers (2024-10-15T02:55:07Z) - CoCPF: Coordinate-based Continuous Projection Field for Ill-Posed Inverse Problem in Imaging [78.734927709231]
Sparse-view computed tomography (SVCT) reconstruction aims to acquire CT images based on sparsely-sampled measurements.
Due to ill-posedness, implicit neural representation (INR) techniques may leave considerable holes'' (i.e., unmodeled spaces) in their fields, leading to sub-optimal results.
We propose the Coordinate-based Continuous Projection Field (CoCPF), which aims to build hole-free representation fields for SVCT reconstruction.
arXiv Detail & Related papers (2024-06-21T08:38:30Z) - 360 Layout Estimation via Orthogonal Planes Disentanglement and Multi-view Geometric Consistency Perception [56.84921040837699]
Existing panoramic layout estimation solutions tend to recover room boundaries from a vertically compressed sequence, yielding imprecise results.
We propose an orthogonal plane disentanglement network (termed DOPNet) to distinguish ambiguous semantics.
We also present an unsupervised adaptation technique tailored for horizon-depth and ratio representations.
Our solution outperforms other SoTA models on both monocular layout estimation and multi-view layout estimation tasks.
arXiv Detail & Related papers (2023-12-26T12:16:03Z) - Handbook on Leveraging Lines for Two-View Relative Pose Estimation [82.72686460985297]
We propose an approach for estimating the relative pose between image pairs by jointly exploiting points, lines, and their coincidences in a hybrid manner.
Our hybrid framework combines the advantages of all configurations, enabling robust and accurate estimation in challenging environments.
arXiv Detail & Related papers (2023-09-27T21:43:04Z) - Contour Context: Abstract Structural Distribution for 3D LiDAR Loop
Detection and Metric Pose Estimation [31.968749056155467]
This paper proposes a simple, effective, and efficient topological loop closure detection pipeline with accurate 3-DoF metric pose estimation.
We interpret the Cartesian birds' eye view (BEV) image projected from 3D LiDAR points as layered distribution of structures.
A retrieval key is designed to accelerate the search of a database indexed by layered KD-trees.
arXiv Detail & Related papers (2023-02-13T07:18:24Z) - 2D LiDAR and Camera Fusion Using Motion Cues for Indoor Layout
Estimation [2.6905021039717987]
A ground robot explores an indoor space with a single floor and vertical walls, and collects a sequence of intensity images and 2D LiDAR datasets.
The alignment of sensor outputs and image segmentation are computed jointly by aligning LiDAR points.
The ambiguity in images for ground-wall boundary extraction is removed with the assistance of LiDAR observations.
arXiv Detail & Related papers (2022-04-24T06:26:02Z) - LED2-Net: Monocular 360 Layout Estimation via Differentiable Depth
Rendering [59.63979143021241]
We formulate the task of 360 layout estimation as a problem of predicting depth on the horizon line of a panorama.
We propose the Differentiable Depth Rendering procedure to make the conversion from layout to depth prediction differentiable.
Our method achieves state-of-the-art performance on numerous 360 layout benchmark datasets.
arXiv Detail & Related papers (2021-04-01T15:48:41Z) - Pixel-Pair Occlusion Relationship Map(P2ORM): Formulation, Inference &
Application [20.63938300312815]
We formalize concepts around geometric occlusion in 2D images (i.e., ignoring semantics)
We propose a novel unified formulation of both occlusion boundaries and occlusion orientations via a pixel-pair occlusion relation.
Experiments on a variety of datasets demonstrate that our method outperforms existing ones on this task.
We also propose a new depth map refinement method that consistently improve the performance of state-of-the-art monocular depth estimation methods.
arXiv Detail & Related papers (2020-07-23T15:52:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.