Connecting the Dots: Floorplan Reconstruction Using Two-Level Queries
- URL: http://arxiv.org/abs/2211.15658v2
- Date: Tue, 28 Mar 2023 02:20:16 GMT
- Title: Connecting the Dots: Floorplan Reconstruction Using Two-Level Queries
- Authors: Yuanwen Yue, Theodora Kontogianni, Konrad Schindler, Francis Engelmann
- Abstract summary: We develop a novel Transformer architecture that generates polygons of multiple rooms in parallel.
Our method achieves a new state-of-the-art for two challenging datasets, Structured3D and SceneCAD.
It can readily be extended to predict additional information, i.e., semantic room types and architectural elements like doors and windows.
- Score: 27.564355569013706
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We address 2D floorplan reconstruction from 3D scans. Existing approaches
typically employ heuristically designed multi-stage pipelines. Instead, we
formulate floorplan reconstruction as a single-stage structured prediction
task: find a variable-size set of polygons, which in turn are variable-length
sequences of ordered vertices. To solve it we develop a novel Transformer
architecture that generates polygons of multiple rooms in parallel, in a
holistic manner without hand-crafted intermediate stages. The model features
two-level queries for polygons and corners, and includes polygon matching to
make the network end-to-end trainable. Our method achieves a new
state-of-the-art for two challenging datasets, Structured3D and SceneCAD, along
with significantly faster inference than previous methods. Moreover, it can
readily be extended to predict additional information, i.e., semantic room
types and architectural elements like doors and windows. Our code and models
are available at: https://github.com/ywyue/RoomFormer.
Related papers
- PolyRoom: Room-aware Transformer for Floorplan Reconstruction [17.154556344393743]
We present PolyRoom, a room-aware Transformer to reconstruct floorplans from point clouds.
Specifically, we adopt a uniform sampling floorplan representation to enable dense supervision during training and effective utilization of angle information.
Results on two widely used datasets demonstrate that PolyRoom surpasses current state-of-the-art methods both quantitatively and qualitatively.
arXiv Detail & Related papers (2024-07-15T04:53:10Z) - ConDaFormer: Disassembled Transformer with Local Structure Enhancement
for 3D Point Cloud Understanding [105.98609765389895]
Transformers have been recently explored for 3D point cloud understanding.
A large number of points, over 0.1 million, make the global self-attention infeasible for point cloud data.
In this paper, we develop a new transformer block, named ConDaFormer.
arXiv Detail & Related papers (2023-12-18T11:19:45Z) - HiT: Building Mapping with Hierarchical Transformers [43.31497052507252]
We propose a simple and novel building mapping method with Hierarchical Transformers, called HiT.
HiT builds on a two-stage detection architecture by adding a polygon head parallel to classification and bounding box regression heads.
Our method achieves a new state-of-the-art in terms of instance segmentation and polygonal metrics compared with state-of-the-art methods.
arXiv Detail & Related papers (2023-09-18T10:24:25Z) - Single-view 3D Mesh Reconstruction for Seen and Unseen Categories [69.29406107513621]
Single-view 3D Mesh Reconstruction is a fundamental computer vision task that aims at recovering 3D shapes from single-view RGB images.
This paper tackles Single-view 3D Mesh Reconstruction, to study the model generalization on unseen categories.
We propose an end-to-end two-stage network, GenMesh, to break the category boundaries in reconstruction.
arXiv Detail & Related papers (2022-08-04T14:13:35Z) - MCTS with Refinement for Proposals Selection Games in Scene
Understanding [32.92475660892122]
We propose a novel method applicable in many scene understanding problems that adapts the Monte Carlo Tree Search (MCTS) algorithm.
From a generated pool of proposals, our method jointly selects and optimize proposals that maximize the objective term.
Our method shows high performance on the Matterport3D dataset without introducing hard constraints on room layout configurations.
arXiv Detail & Related papers (2022-07-07T10:15:54Z) - Neural Template: Topology-aware Reconstruction and Disentangled
Generation of 3D Meshes [52.038346313823524]
This paper introduces a novel framework called DTNet for 3D mesh reconstruction and generation via Disentangled Topology.
Our method is able to produce high-quality meshes, particularly with diverse topologies, as compared with the state-of-the-art methods.
arXiv Detail & Related papers (2022-06-10T08:32:57Z) - Neural 3D Scene Reconstruction with the Manhattan-world Assumption [58.90559966227361]
This paper addresses the challenge of reconstructing 3D indoor scenes from multi-view images.
Planar constraints can be conveniently integrated into the recent implicit neural representation-based reconstruction methods.
The proposed method outperforms previous methods by a large margin on 3D reconstruction quality.
arXiv Detail & Related papers (2022-05-05T17:59:55Z) - A Scalable Combinatorial Solver for Elastic Geometrically Consistent 3D
Shape Matching [69.14632473279651]
We present a scalable algorithm for globally optimizing over the space of geometrically consistent mappings between 3D shapes.
We propose a novel primal coupled with a Lagrange dual problem that is several orders of magnitudes faster than previous solvers.
arXiv Detail & Related papers (2022-04-27T09:47:47Z) - Automated LoD-2 Model Reconstruction from Very-HighResolution
Satellite-derived Digital Surface Model and Orthophoto [1.2691047660244335]
We propose a model-driven method that reconstructs LoD-2 building models following a "decomposition-optimization-fitting" paradigm.
Our proposed method has addressed a few technical caveats over existing methods, resulting in practically high-quality results.
arXiv Detail & Related papers (2021-09-08T19:03:09Z) - Implicit Functions in Feature Space for 3D Shape Reconstruction and
Completion [53.885984328273686]
Implicit Feature Networks (IF-Nets) deliver continuous outputs, can handle multiple topologies, and complete shapes for missing or sparse input data.
IF-Nets clearly outperform prior work in 3D object reconstruction in ShapeNet, and obtain significantly more accurate 3D human reconstructions.
arXiv Detail & Related papers (2020-03-03T11:14:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.