Semantic Implicit Neural Scene Representations With Semi-Supervised
Training
- URL: http://arxiv.org/abs/2003.12673v2
- Date: Sun, 17 Jan 2021 01:53:26 GMT
- Title: Semantic Implicit Neural Scene Representations With Semi-Supervised
Training
- Authors: Amit Kohli, Vincent Sitzmann, Gordon Wetzstein
- Abstract summary: We show that implicit neural scene representations can be leveraged to perform per-point semantic segmentation.
Our method is simple, general, and only requires a few tens of labeled 2D segmentation masks.
We explore two novel applications for this semantically aware implicit neural scene representation.
- Score: 47.61092265963234
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The recent success of implicit neural scene representations has presented a
viable new method for how we capture and store 3D scenes. Unlike conventional
3D representations, such as point clouds, which explicitly store scene
properties in discrete, localized units, these implicit representations encode
a scene in the weights of a neural network which can be queried at any
coordinate to produce these same scene properties. Thus far, implicit
representations have primarily been optimized to estimate only the appearance
and/or 3D geometry information in a scene. We take the next step and
demonstrate that an existing implicit representation (SRNs) is actually
multi-modal; it can be further leveraged to perform per-point semantic
segmentation while retaining its ability to represent appearance and geometry.
To achieve this multi-modal behavior, we utilize a semi-supervised learning
strategy atop the existing pre-trained scene representation. Our method is
simple, general, and only requires a few tens of labeled 2D segmentation masks
in order to achieve dense 3D semantic segmentation. We explore two novel
applications for this semantically aware implicit neural scene representation:
3D novel view and semantic label synthesis given only a single input RGB image
or 2D label mask, as well as 3D interpolation of appearance and semantics.
Related papers
- MOSE: Monocular Semantic Reconstruction Using NeRF-Lifted Noisy Priors [11.118490283303407]
We propose a neural field semantic reconstruction approach to lift inferred image-level noisy priors to 3D.
Our method produces accurate semantics and geometry in both 3D and 2D space.
arXiv Detail & Related papers (2024-09-21T05:12:13Z) - Efficient 3D Instance Mapping and Localization with Neural Fields [39.73128916618561]
We tackle the problem of learning an implicit scene representation for 3D instance segmentation from a sequence of posed RGB images.
We introduce 3DIML, a novel framework that efficiently learns a neural label field which can render 3D instance segmentation masks from novel viewpoints.
arXiv Detail & Related papers (2024-03-28T19:25:25Z) - GNeSF: Generalizable Neural Semantic Fields [48.49860868061573]
We introduce a generalizable 3D segmentation framework based on implicit representation.
We propose a novel soft voting mechanism to aggregate the 2D semantic information from different views for each 3D point.
Our approach can even outperform existing strong supervision-based approaches with only 2D annotations.
arXiv Detail & Related papers (2023-10-24T10:40:51Z) - SceneDreamer: Unbounded 3D Scene Generation from 2D Image Collections [49.802462165826554]
We present SceneDreamer, an unconditional generative model for unbounded 3D scenes.
Our framework is learned from in-the-wild 2D image collections only, without any 3D annotations.
arXiv Detail & Related papers (2023-02-02T18:59:16Z) - Neural Feature Fusion Fields: 3D Distillation of Self-Supervised 2D
Image Representations [92.88108411154255]
We present a method that improves dense 2D image feature extractors when the latter are applied to the analysis of multiple images reconstructible as a 3D scene.
We show that our method not only enables semantic understanding in the context of scene-specific neural fields without the use of manual labels, but also consistently improves over the self-supervised 2D baselines.
arXiv Detail & Related papers (2022-09-07T23:24:09Z) - CompNVS: Novel View Synthesis with Scene Completion [83.19663671794596]
We propose a generative pipeline performing on a sparse grid-based neural scene representation to complete unobserved scene parts.
We process encoded image features in 3D space with a geometry completion network and a subsequent texture inpainting network to extrapolate the missing area.
Photorealistic image sequences can be finally obtained via consistency-relevant differentiable rendering.
arXiv Detail & Related papers (2022-07-23T09:03:13Z) - Neural Groundplans: Persistent Neural Scene Representations from a
Single Image [90.04272671464238]
We present a method to map 2D image observations of a scene to a persistent 3D scene representation.
We propose conditional neural groundplans as persistent and memory-efficient scene representations.
arXiv Detail & Related papers (2022-07-22T17:41:24Z) - NeSF: Neural Semantic Fields for Generalizable Semantic Segmentation of
3D Scenes [25.26518805603798]
NeSF is a method for producing 3D semantic fields from posed RGB images alone.
Our method is the first to offer truly dense 3D scene segmentations requiring only 2D supervision for training.
arXiv Detail & Related papers (2021-11-25T21:44:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.