Walk2Map: Extracting Floor Plans from Indoor Walk Trajectories
- URL: http://arxiv.org/abs/2103.00262v1
- Date: Sat, 27 Feb 2021 16:29:09 GMT
- Title: Walk2Map: Extracting Floor Plans from Indoor Walk Trajectories
- Authors: Claudio Mura, Renato Pajarola, Konrad Schindler, Niloy Mitra
- Abstract summary: We present Walk2Map, a data-driven approach to generate floor plans from trajectories of a person walking inside the rooms.
Thanks to advances in data-driven inertial odometry, such minimalistic input data can be acquired from the IMU readings of consumer-level smartphones.
We train our networks using scanned 3D indoor models and apply them in a cascaded fashion on an indoor walk trajectory.
- Score: 23.314557741879664
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent years have seen a proliferation of new digital products for the
efficient management of indoor spaces, with important applications like
emergency management, virtual property showcasing and interior design. These
products rely on accurate 3D models of the environments considered, including
information on both architectural and non-permanent elements. These models must
be created from measured data such as RGB-D images or 3D point clouds, whose
capture and consolidation involves lengthy data workflows. This strongly limits
the rate at which 3D models can be produced, preventing the adoption of many
digital services for indoor space management. We provide an alternative to such
data-intensive procedures by presenting Walk2Map, a data-driven approach to
generate floor plans only from trajectories of a person walking inside the
rooms. Thanks to recent advances in data-driven inertial odometry, such
minimalistic input data can be acquired from the IMU readings of consumer-level
smartphones, which allows for an effortless and scalable mapping of real-world
indoor spaces. Our work is based on learning the latent relation between an
indoor walk trajectory and the information represented in a floor plan:
interior space footprint, portals, and furniture. We distinguish between
recovering area-related (interior footprint, furniture) and wall-related
(doors) information and use two different neural architectures for the two
tasks: an image-based Encoder-Decoder and a Graph Convolutional Network,
respectively. We train our networks using scanned 3D indoor models and apply
them in a cascaded fashion on an indoor walk trajectory at inference time. We
perform a qualitative and quantitative evaluation using both simulated and
measured, real-world trajectories, and compare against a baseline method for
image-to-image translation. The experiments confirm the feasibility of our
approach.
Related papers
- Syn-to-Real Unsupervised Domain Adaptation for Indoor 3D Object Detection [50.448520056844885]
We propose a novel framework for syn-to-real unsupervised domain adaptation in indoor 3D object detection.
Our adaptation results from synthetic dataset 3D-FRONT to real-world datasets ScanNetV2 and SUN RGB-D demonstrate remarkable mAP25 improvements of 9.7% and 9.1% over Source-Only baselines.
arXiv Detail & Related papers (2024-06-17T08:18:41Z) - Indoor and Outdoor 3D Scene Graph Generation via Language-Enabled Spatial Ontologies [16.396336068724484]
This paper proposes an approach to build 3D scene graphs in arbitrary indoor and outdoor environments.
The hierarchy of concepts that describe an outdoor environment is more complex than for indoors.
The lack of training data prevents the straightforward application of learning-based tools used in indoor settings.
arXiv Detail & Related papers (2023-12-18T21:20:28Z) - AutoDecoding Latent 3D Diffusion Models [95.7279510847827]
We present a novel approach to the generation of static and articulated 3D assets that has a 3D autodecoder at its core.
The 3D autodecoder framework embeds properties learned from the target dataset in the latent space.
We then identify the appropriate intermediate volumetric latent space, and introduce robust normalization and de-normalization operations.
arXiv Detail & Related papers (2023-07-07T17:59:14Z) - 3D Data Augmentation for Driving Scenes on Camera [50.41413053812315]
We propose a 3D data augmentation approach termed Drive-3DAug, aiming at augmenting the driving scenes on camera in the 3D space.
We first utilize Neural Radiance Field (NeRF) to reconstruct the 3D models of background and foreground objects.
Then, augmented driving scenes can be obtained by placing the 3D objects with adapted location and orientation at the pre-defined valid region of backgrounds.
arXiv Detail & Related papers (2023-03-18T05:51:05Z) - Disentangling Variational Autoencoders [0.0]
A variational autoencoder (VAE) projects an input set of high-dimensional data to a lower-dimensional, latent space.
We implement three different VAE models from the literature and train them on a dataset of 60,000 images of hand-written digits.
We investigate the trade-offs between the quality of the reconstruction of the decoded images and the level of disentanglement of the latent space.
arXiv Detail & Related papers (2022-11-14T19:22:41Z) - Embodied Navigation at the Art Gallery [43.52107532692226]
We build and release a new 3D space with unique characteristics: the one of a complete art museum.
Compared with existing 3D scenes, the collected space is ampler, richer in visual features, and provides very sparse occupancy information.
We deliver a new benchmark for PointGoal navigation inside this new space.
arXiv Detail & Related papers (2022-04-19T18:00:06Z) - Simple and Effective Synthesis of Indoor 3D Scenes [78.95697556834536]
We study the problem of immersive 3D indoor scenes from one or more images.
Our aim is to generate high-resolution images and videos from novel viewpoints.
We propose an image-to-image GAN that maps directly from reprojections of incomplete point clouds to full high-resolution RGB-D images.
arXiv Detail & Related papers (2022-04-06T17:54:46Z) - SimIPU: Simple 2D Image and 3D Point Cloud Unsupervised Pre-Training for
Spatial-Aware Visual Representations [85.38562724999898]
We propose a 2D Image and 3D Point cloud Unsupervised pre-training strategy, called SimIPU.
Specifically, we develop a multi-modal contrastive learning framework that consists of an intra-modal spatial perception module and an inter-modal feature interaction module.
To the best of our knowledge, this is the first study to explore contrastive learning pre-training strategies for outdoor multi-modal datasets.
arXiv Detail & Related papers (2021-12-09T03:27:00Z) - Cylindrical and Asymmetrical 3D Convolution Networks for LiDAR-based
Perception [122.53774221136193]
State-of-the-art methods for driving-scene LiDAR-based perception often project the point clouds to 2D space and then process them via 2D convolution.
A natural remedy is to utilize the 3D voxelization and 3D convolution network.
We propose a new framework for the outdoor LiDAR segmentation, where cylindrical partition and asymmetrical 3D convolution networks are designed to explore the 3D geometric pattern.
arXiv Detail & Related papers (2021-09-12T06:25:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.