General 3D Room Layout from a Single View by Render-and-Compare
- URL: http://arxiv.org/abs/2001.02149v2
- Date: Tue, 21 Jul 2020 15:41:56 GMT
- Title: General 3D Room Layout from a Single View by Render-and-Compare
- Authors: Sinisa Stekovic, Shreyas Hampali, Mahdi Rad, Sayan Deb Sarkar,
Friedrich Fraundorfer, Vincent Lepetit
- Abstract summary: We present a novel method to reconstruct the 3D layout of a room from a single perspective view.
Our dataset consists of 293 images from ScanNet, which we annotated with precise 3D layouts.
- Score: 36.94817376590415
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a novel method to reconstruct the 3D layout of a room (walls,
floors, ceilings) from a single perspective view in challenging conditions, by
contrast with previous single-view methods restricted to cuboid-shaped layouts.
This input view can consist of a color image only, but considering a depth map
results in a more accurate reconstruction. Our approach is formalized as
solving a constrained discrete optimization problem to find the set of 3D
polygons that constitute the layout. In order to deal with occlusions between
components of the layout, which is a problem ignored by previous works, we
introduce an analysis-by-synthesis method to iteratively refine the 3D layout
estimate. As no dataset was available to evaluate our method quantitatively, we
created one together with several appropriate metrics. Our dataset consists of
293 images from ScanNet, which we annotated with precise 3D layouts. It offers
three times more samples than the popular NYUv2 303 benchmark, and a much
larger variety of layouts.
Related papers
- High-fidelity 3D GAN Inversion by Pseudo-multi-view Optimization [51.878078860524795]
We present a high-fidelity 3D generative adversarial network (GAN) inversion framework that can synthesize photo-realistic novel views.
Our approach enables high-fidelity 3D rendering from a single image, which is promising for various applications of AI-generated 3D content.
arXiv Detail & Related papers (2022-11-28T18:59:52Z) - Towards High-Fidelity Single-view Holistic Reconstruction of Indoor
Scenes [50.317223783035075]
We present a new framework to reconstruct holistic 3D indoor scenes from single-view images.
We propose an instance-aligned implicit function (InstPIFu) for detailed object reconstruction.
Our code and model will be made publicly available.
arXiv Detail & Related papers (2022-07-18T14:54:57Z) - MCTS with Refinement for Proposals Selection Games in Scene
Understanding [32.92475660892122]
We propose a novel method applicable in many scene understanding problems that adapts the Monte Carlo Tree Search (MCTS) algorithm.
From a generated pool of proposals, our method jointly selects and optimize proposals that maximize the objective term.
Our method shows high performance on the Matterport3D dataset without introducing hard constraints on room layout configurations.
arXiv Detail & Related papers (2022-07-07T10:15:54Z) - Neural 3D Scene Reconstruction with the Manhattan-world Assumption [58.90559966227361]
This paper addresses the challenge of reconstructing 3D indoor scenes from multi-view images.
Planar constraints can be conveniently integrated into the recent implicit neural representation-based reconstruction methods.
The proposed method outperforms previous methods by a large margin on 3D reconstruction quality.
arXiv Detail & Related papers (2022-05-05T17:59:55Z) - Learning to Reconstruct 3D Non-Cuboid Room Layout from a Single RGB
Image [32.5277483805739]
Single-image room layout reconstruction aims to reconstruct the enclosed 3D structure of a room from a single image.
This paper considers a more general indoor assumption, i.e., the room layout consists of a single ceiling, a single floor, and several vertical walls.
arXiv Detail & Related papers (2021-04-16T09:24:08Z) - LED2-Net: Monocular 360 Layout Estimation via Differentiable Depth
Rendering [59.63979143021241]
We formulate the task of 360 layout estimation as a problem of predicting depth on the horizon line of a panorama.
We propose the Differentiable Depth Rendering procedure to make the conversion from layout to depth prediction differentiable.
Our method achieves state-of-the-art performance on numerous 360 layout benchmark datasets.
arXiv Detail & Related papers (2021-04-01T15:48:41Z) - A Divide et Impera Approach for 3D Shape Reconstruction from Multiple
Views [49.03830902235915]
Estimating the 3D shape of an object from a single or multiple images has gained popularity thanks to the recent breakthroughs powered by deep learning.
This paper proposes to rely on viewpoint variant reconstructions by merging the visible information from the given views.
To validate the proposed method, we perform a comprehensive evaluation on the ShapeNet reference benchmark in terms of relative pose estimation and 3D shape reconstruction.
arXiv Detail & Related papers (2020-11-17T09:59:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.