SSLayout360: Semi-Supervised Indoor Layout Estimation from 360-Degree
Panorama
- URL: http://arxiv.org/abs/2103.13696v2
- Date: Mon, 29 Mar 2021 06:22:43 GMT
- Title: SSLayout360: Semi-Supervised Indoor Layout Estimation from 360-Degree
Panorama
- Authors: Phi Vu Tran
- Abstract summary: We propose the first approach to learn representations of room corners and boundaries by using a combination of labeled and unlabeled data.
Our approach can advance layout estimation of complex indoor scenes using as few as 20 labeled examples.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Recent years have seen flourishing research on both semi-supervised learning
and 3D room layout reconstruction. In this work, we explore the intersection of
these two fields to advance the research objective of enabling more accurate 3D
indoor scene modeling with less labeled data. We propose the first approach to
learn representations of room corners and boundaries by using a combination of
labeled and unlabeled data for improved layout estimation in a 360-degree
panoramic scene. Through extensive comparative experiments, we demonstrate that
our approach can advance layout estimation of complex indoor scenes using as
few as 20 labeled examples. When coupled with a layout predictor pre-trained on
synthetic data, our semi-supervised method matches the fully supervised
counterpart using only 12% of the labels. Our work takes an important first
step towards robust semi-supervised layout estimation that can enable many
applications in 3D perception with limited labeled data.
Related papers
- Dual-Perspective Knowledge Enrichment for Semi-Supervised 3D Object
Detection [55.210991151015534]
We present a novel Dual-Perspective Knowledge Enrichment approach named DPKE for semi-supervised 3D object detection.
Our DPKE enriches the knowledge of limited training data, particularly unlabeled data, from two perspectives: data-perspective and feature-perspective.
arXiv Detail & Related papers (2024-01-10T08:56:07Z) - 360-MLC: Multi-view Layout Consistency for Self-training and
Hyper-parameter Tuning [40.93848397359068]
We present 360-MLC, a self-training method based on multi-view layout consistency for finetuning monocular room- models.
We leverage the entropy information in multiple layout estimations as a quantitative metric to measure the geometry consistency of the scene.
arXiv Detail & Related papers (2022-10-24T03:31:48Z) - Semi-Supervised Single-View 3D Reconstruction via Prototype Shape Priors [79.80916315953374]
We propose SSP3D, a semi-supervised framework for 3D reconstruction.
We introduce an attention-guided prototype shape prior module for guiding realistic object reconstruction.
Our approach also performs well when transferring to real-world Pix3D datasets under labeling ratios of 10%.
arXiv Detail & Related papers (2022-09-30T11:19:25Z) - Monocular Spherical Depth Estimation with Explicitly Connected Weak
Layout Cues [27.15511982413305]
We generate a geometric vision (360V) dataset that includes multiple modalities, multi-view stereo data and automatically generated weak layout cues.
We rely on depth-based layout reconstruction and layout-based depth attention, demonstrating increased performance across both tasks.
By using single 360 cameras to scan rooms, the opportunity for facile and quick building-scale 3D scanning arises.
arXiv Detail & Related papers (2022-06-22T20:10:45Z) - 360-DFPE: Leveraging Monocular 360-Layouts for Direct Floor Plan
Estimation [43.56963653723287]
We present 360-DFPE, a sequential floor plan estimation method that directly takes 360-images as input without relying on active sensors or 3D information.
Our results show that our monocular solution achieves favorable performance against the current state-of-the-art algorithms.
arXiv Detail & Related papers (2021-12-12T08:36:41Z) - LED2-Net: Monocular 360 Layout Estimation via Differentiable Depth
Rendering [59.63979143021241]
We formulate the task of 360 layout estimation as a problem of predicting depth on the horizon line of a panorama.
We propose the Differentiable Depth Rendering procedure to make the conversion from layout to depth prediction differentiable.
Our method achieves state-of-the-art performance on numerous 360 layout benchmark datasets.
arXiv Detail & Related papers (2021-04-01T15:48:41Z) - Semi-Automatic Data Annotation guided by Feature Space Projection [117.9296191012968]
We present a semi-automatic data annotation approach based on suitable feature space projection and semi-supervised label estimation.
We validate our method on the popular MNIST dataset and on images of human intestinal parasites with and without fecal impurities.
Our results demonstrate the added-value of visual analytics tools that combine complementary abilities of humans and machines for more effective machine learning.
arXiv Detail & Related papers (2020-07-27T17:03:50Z) - LayoutMP3D: Layout Annotation of Matterport3D [59.11106101006007]
We consider the Matterport3D dataset with their originally provided depth map ground truths and further release our annotations for layout ground truths from a subset of Matterport3D.
Our dataset provides both the layout and depth information, which enables the opportunity to explore the environment by integrating both cues.
arXiv Detail & Related papers (2020-03-30T14:40:56Z) - Scan2Plan: Efficient Floorplan Generation from 3D Scans of Indoor Scenes [9.71137838903781]
Scan2Plan is a novel approach for accurate estimation of a floorplan from a 3D scan of the structural elements of indoor environments.
The proposed method incorporates a two-stage approach where the initial stage clusters an unordered point cloud representation of the scene.
The subsequent stage estimates a closed perimeter, parameterized by a simple polygon, for each individual room.
The final floorplan is simply an assembly of all such room perimeters in the global co-ordinate system.
arXiv Detail & Related papers (2020-03-16T17:59:41Z) - Learning multiview 3D point cloud registration [74.39499501822682]
We present a novel, end-to-end learnable, multiview 3D point cloud registration algorithm.
Our approach outperforms the state-of-the-art by a significant margin, while being end-to-end trainable and computationally less costly.
arXiv Detail & Related papers (2020-01-15T03:42:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.