Floorplan-Aware Camera Poses Refinement
- URL: http://arxiv.org/abs/2210.04572v1
- Date: Mon, 10 Oct 2022 11:24:10 GMT
- Title: Floorplan-Aware Camera Poses Refinement
- Authors: Anna Sokolova, Filipp Nikitin, Anna Vorontsova, Anton Konushin
- Abstract summary: We argue that a floorplan is a useful source of spatial information, which can guide a 3D model optimization.
We propose a novel optimization algorithm expanding conventional BA that leverages the prior knowledge about the scene structure in the form of a floorplan.
Our experiments on the Redwood dataset and our self-captured data demonstrate that utilizing floorplan improves accuracy of 3D reconstructions.
- Score: 2.294014185517203
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Processing large indoor scenes is a challenging task, as scan registration
and camera trajectory estimation methods accumulate errors across time. As a
result, the quality of reconstructed scans is insufficient for some
applications, such as visual-based localization and navigation, where the
correct position of walls is crucial.
For many indoor scenes, there exists an image of a technical floorplan that
contains information about the geometry and main structural elements of the
scene, such as walls, partitions, and doors. We argue that such a floorplan is
a useful source of spatial information, which can guide a 3D model
optimization.
The standard RGB-D 3D reconstruction pipeline consists of a tracking module
applied to an RGB-D sequence and a bundle adjustment (BA) module that takes the
posed RGB-D sequence and corrects the camera poses to improve consistency. We
propose a novel optimization algorithm expanding conventional BA that leverages
the prior knowledge about the scene structure in the form of a floorplan. Our
experiments on the Redwood dataset and our self-captured data demonstrate that
utilizing floorplan improves accuracy of 3D reconstructions.
Related papers
- SCRREAM : SCan, Register, REnder And Map:A Framework for Annotating Accurate and Dense 3D Indoor Scenes with a Benchmark [43.88114765730359]
SCRREAM allows annotation of fully dense meshes of objects in the scene and registers camera poses on the real image sequence.
We show the details of the dataset annotation pipeline and showcase four possible variants of datasets.
Recent pipelines for indoor reconstruction and SLAM serve as new benchmarks.
arXiv Detail & Related papers (2024-10-30T05:53:07Z) - MM3DGS SLAM: Multi-modal 3D Gaussian Splatting for SLAM Using Vision, Depth, and Inertial Measurements [59.70107451308687]
We show for the first time that using 3D Gaussians for map representation with unposed camera images and inertial measurements can enable accurate SLAM.
Our method, MM3DGS, addresses the limitations of prior rendering by enabling faster scale awareness, and improved trajectory tracking.
We also release a multi-modal dataset, UT-MM, collected from a mobile robot equipped with a camera and an inertial measurement unit.
arXiv Detail & Related papers (2024-04-01T04:57:41Z) - Multi-Modal Dataset Acquisition for Photometrically Challenging Object [56.30027922063559]
This paper addresses the limitations of current datasets for 3D vision tasks in terms of accuracy, size, realism, and suitable imaging modalities for photometrically challenging objects.
We propose a novel annotation and acquisition pipeline that enhances existing 3D perception and 6D object pose datasets.
arXiv Detail & Related papers (2023-08-21T10:38:32Z) - Lazy Visual Localization via Motion Averaging [89.8709956317671]
We show that it is possible to achieve high localization accuracy without reconstructing the scene from the database.
Experiments show that our visual localization proposal, LazyLoc, achieves comparable performance against state-of-the-art structure-based methods.
arXiv Detail & Related papers (2023-07-19T13:40:45Z) - 3D Reconstruction of Spherical Images based on Incremental Structure
from Motion [2.6432771146480283]
This study investigates the algorithms for the relative orientation using spherical correspondences, absolute orientation using 3D correspondences between scene and spherical points, and the cost functions for BA (bundle adjustment) optimization.
An incremental SfM (Structure from Motion) workflow has been proposed for spherical images using the above-mentioned algorithms.
arXiv Detail & Related papers (2023-06-22T09:49:28Z) - StructuredMesh: 3D Structured Optimization of Fa\c{c}ade Components on
Photogrammetric Mesh Models using Binary Integer Programming [17.985961236568663]
We present StructuredMesh, a novel approach for reconstructing faccade structures conforming to the regularity of buildings within photogrammetric mesh models.
Our method involves capturing multi-view color and depth images of the building model using a virtual camera.
We then utilize the depth image to remap these boxes into 3D space, generating an initial faccade layout.
arXiv Detail & Related papers (2023-06-07T06:40:54Z) - BS3D: Building-scale 3D Reconstruction from RGB-D Images [25.604775584883413]
We propose an easy-to-use framework for acquiring building-scale 3D reconstruction using a consumer depth camera.
Unlike complex and expensive acquisition setups, our system enables crowd-sourcing, which can greatly benefit data-hungry algorithms.
arXiv Detail & Related papers (2023-01-03T11:46:14Z) - Towards Non-Line-of-Sight Photography [48.491977359971855]
Non-line-of-sight (NLOS) imaging is based on capturing the multi-bounce indirect reflections from the hidden objects.
Active NLOS imaging systems rely on the capture of the time of flight of light through the scene.
We propose a new problem formulation, called NLOS photography, to specifically address this deficiency.
arXiv Detail & Related papers (2021-09-16T08:07:13Z) - Plan2Scene: Converting Floorplans to 3D Scenes [36.34298107648571]
We address the task of converting a floorplan and a set of associated photos of a residence into a textured 3D mesh model.
Our system 1) lifts a floorplan image to a 3D mesh model; 2) synthesizes surface textures based on the input photos; and 3) infers textures for unobserved surfaces using a graph neural network architecture.
arXiv Detail & Related papers (2021-06-09T20:32:20Z) - LED2-Net: Monocular 360 Layout Estimation via Differentiable Depth
Rendering [59.63979143021241]
We formulate the task of 360 layout estimation as a problem of predicting depth on the horizon line of a panorama.
We propose the Differentiable Depth Rendering procedure to make the conversion from layout to depth prediction differentiable.
Our method achieves state-of-the-art performance on numerous 360 layout benchmark datasets.
arXiv Detail & Related papers (2021-04-01T15:48:41Z) - Lightweight Multi-View 3D Pose Estimation through Camera-Disentangled
Representation [57.11299763566534]
We present a solution to recover 3D pose from multi-view images captured with spatially calibrated cameras.
We exploit 3D geometry to fuse input images into a unified latent representation of pose, which is disentangled from camera view-points.
Our architecture then conditions the learned representation on camera projection operators to produce accurate per-view 2d detections.
arXiv Detail & Related papers (2020-04-05T12:52:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.