FloorGenT: Generative Vector Graphic Model of Floor Plans for Robotics
- URL: http://arxiv.org/abs/2203.03385v1
- Date: Mon, 7 Mar 2022 13:42:48 GMT
- Title: FloorGenT: Generative Vector Graphic Model of Floor Plans for Robotics
- Authors: Ludvig Ericson, Patric Jensfelt
- Abstract summary: We show that by modelling floor plans as sequences of line segments seen from a particular point of view, recent advances in autoregressive sequence modelling can be leveraged to model and predict floor plans.
- Score: 5.71097144710995
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Floor plans are the basis of reasoning in and communicating about indoor
environments. In this paper, we show that by modelling floor plans as sequences
of line segments seen from a particular point of view, recent advances in
autoregressive sequence modelling can be leveraged to model and predict floor
plans. The line segments are canonicalized and translated to sequence of tokens
and an attention-based neural network is used to fit a one-step distribution
over next tokens. We fit the network to sequences derived from a set of
large-scale floor plans, and demonstrate the capabilities of the model in four
scenarios: novel floor plan generation, completion of partially observed floor
plans, generation of floor plans from simulated sensor data, and finally, the
applicability of a floor plan model in predicting the shortest distance with
partial knowledge of the environment.
Related papers
- DStruct2Design: Data and Benchmarks for Data Structure Driven Generative Floor Plan Design [5.567585193148804]
We construct a new dataset for this data-structure to data-structure formulation of floorplan generation.
We explore the task of floorplan generation given a partial or complete set of constraints.
We demonstrate the feasibility of using floorplan data structure conditioned LLMs for the problem of floorplan generation respecting numerical constraints.
arXiv Detail & Related papers (2024-07-22T15:27:55Z) - Planning as In-Painting: A Diffusion-Based Embodied Task Planning
Framework for Environments under Uncertainty [56.30846158280031]
Task planning for embodied AI has been one of the most challenging problems.
We propose a task-agnostic method named 'planning as in-painting'
The proposed framework achieves promising performances in various embodied AI tasks.
arXiv Detail & Related papers (2023-12-02T10:07:17Z) - Compositional Foundation Models for Hierarchical Planning [52.18904315515153]
We propose a foundation model which leverages expert foundation model trained on language, vision and action data individually together to solve long-horizon tasks.
We use a large language model to construct symbolic plans that are grounded in the environment through a large video diffusion model.
Generated video plans are then grounded to visual-motor control, through an inverse dynamics model that infers actions from generated videos.
arXiv Detail & Related papers (2023-09-15T17:44:05Z) - 2D Floor Plan Segmentation Based on Down-sampling [1.4502611532302039]
We propose a novel 2D floor plan segmentation technique based on a down-sampling approach.
Our method employs continuous down-sampling on a floor plan to maintain its structural information while reducing its complexity.
arXiv Detail & Related papers (2023-03-24T04:39:50Z) - End-to-end Graph-constrained Vectorized Floorplan Generation with
Panoptic Refinement [16.103152098205566]
We aim to synthesize floorplans as sequences of 1-D vectors, which eases user interaction and design customization.
In the first stage, we encode the room connectivity graph input by users with a graphal network (GCN), then apply an autoregressive transformer network to generate an initial floorplan sequence.
To polish the initial design and generate more visually appealing floorplans, we further propose a novel panoptic refinement network(PRN) composed of a GCN and a transformer network.
arXiv Detail & Related papers (2022-07-27T03:19:20Z) - Visual Learning-based Planning for Continuous High-Dimensional POMDPs [81.16442127503517]
Visual Tree Search (VTS) is a learning and planning procedure that combines generative models learned offline with online model-based POMDP planning.
VTS bridges offline model training and online planning by utilizing a set of deep generative observation models to predict and evaluate the likelihood of image observations in a Monte Carlo tree search planner.
We show that VTS is robust to different observation noises and, since it utilizes online, model-based planning, can adapt to different reward structures without the need to re-train.
arXiv Detail & Related papers (2021-12-17T11:53:31Z) - Graph-Based Generative Representation Learning of Semantically and
Behaviorally Augmented Floorplans [12.488287536032747]
We present a floorplan embedding technique that uses an attributed graph to represent the geometric information as well as design semantics and behavioral features of the inhabitants as node and edge attributes.
A Long Short-Term Memory (LSTM) Variational Autoencoder (VAE) architecture is proposed and trained to embed attributed graphs as vectors in a continuous space.
A user study is conducted to evaluate the coupling of similar floorplans retrieved from the embedding space with respect to a given input.
arXiv Detail & Related papers (2020-12-08T20:51:56Z) - Long-Horizon Visual Planning with Goal-Conditioned Hierarchical
Predictors [124.30562402952319]
The ability to predict and plan into the future is fundamental for agents acting in the world.
Current learning approaches for visual prediction and planning fail on long-horizon tasks.
We propose a framework for visual prediction and planning that is able to overcome both of these limitations.
arXiv Detail & Related papers (2020-06-23T17:58:56Z) - Divide-and-Conquer Monte Carlo Tree Search For Goal-Directed Planning [78.65083326918351]
We consider alternatives to an implicit sequential planning assumption.
We propose Divide-and-Conquer Monte Carlo Tree Search (DC-MCTS) for approximating the optimal plan.
We show that this algorithmic flexibility over planning order leads to improved results in navigation tasks in grid-worlds.
arXiv Detail & Related papers (2020-04-23T18:08:58Z) - Hallucinative Topological Memory for Zero-Shot Visual Planning [86.20780756832502]
In visual planning (VP), an agent learns to plan goal-directed behavior from observations of a dynamical system obtained offline.
Most previous works on VP approached the problem by planning in a learned latent space, resulting in low-quality visual plans.
Here, we propose a simple VP method that plans directly in image space and displays competitive performance.
arXiv Detail & Related papers (2020-02-27T18:54:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.