Graph-to-3D: End-to-End Generation and Manipulation of 3D Scenes Using
Scene Graphs
- URL: http://arxiv.org/abs/2108.08841v1
- Date: Thu, 19 Aug 2021 17:59:07 GMT
- Title: Graph-to-3D: End-to-End Generation and Manipulation of 3D Scenes Using
Scene Graphs
- Authors: Helisa Dhamo, Fabian Manhardt, Nassir Navab, Federico Tombari
- Abstract summary: Controllable scene synthesis consists of generating 3D information that satisfy underlying specifications.
Scene graphs are representations of a scene composed of objects (nodes) and inter-object relationships (edges)
We propose the first work that directly generates shapes from a scene graph in an end-to-end manner.
- Score: 85.54212143154986
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Controllable scene synthesis consists of generating 3D information that
satisfy underlying specifications. Thereby, these specifications should be
abstract, i.e. allowing easy user interaction, whilst providing enough
interface for detailed control. Scene graphs are representations of a scene,
composed of objects (nodes) and inter-object relationships (edges), proven to
be particularly suited for this task, as they allow for semantic control on the
generated content. Previous works tackling this task often rely on synthetic
data, and retrieve object meshes, which naturally limits the generation
capabilities. To circumvent this issue, we instead propose the first work that
directly generates shapes from a scene graph in an end-to-end manner. In
addition, we show that the same model supports scene modification, using the
respective scene graph as interface. Leveraging Graph Convolutional Networks
(GCN) we train a variational Auto-Encoder on top of the object and edge
categories, as well as 3D shapes and scene layouts, allowing latter sampling of
new scenes and shapes.
Related papers
- GraphDreamer: Compositional 3D Scene Synthesis from Scene Graphs [74.98581417902201]
We propose a novel framework to generate compositional 3D scenes from scene graphs.
By exploiting node and edge information in scene graphs, our method makes better use of the pretrained text-to-image diffusion model.
We conduct both qualitative and quantitative experiments to validate the effectiveness of GraphDreamer.
arXiv Detail & Related papers (2023-11-30T18:59:58Z) - CommonScenes: Generating Commonsense 3D Indoor Scenes with Scene Graph
Diffusion [83.30168660888913]
We present CommonScenes, a fully generative model that converts scene graphs into corresponding controllable 3D scenes.
Our pipeline consists of two branches, one predicting the overall scene layout via a variational auto-encoder and the other generating compatible shapes.
The generated scenes can be manipulated by editing the input scene graph and sampling the noise in the diffusion model.
arXiv Detail & Related papers (2023-05-25T17:39:13Z) - Unconditional Scene Graph Generation [72.53624470737712]
We develop a deep auto-regressive model called SceneGraphGen which can learn the probability distribution over labelled and directed graphs.
We show that the scene graphs generated by SceneGraphGen are diverse and follow the semantic patterns of real-world scenes.
arXiv Detail & Related papers (2021-08-12T17:57:16Z) - SceneGraphFusion: Incremental 3D Scene Graph Prediction from RGB-D
Sequences [76.28527350263012]
We propose a method to incrementally build up semantic scene graphs from a 3D environment given a sequence of RGB-D frames.
We aggregate PointNet features from primitive scene components by means of a graph neural network.
Our approach outperforms 3D scene graph prediction methods by a large margin and its accuracy is on par with other 3D semantic and panoptic segmentation methods while running at 35 Hz.
arXiv Detail & Related papers (2021-03-27T13:00:36Z) - Learning 3D Semantic Scene Graphs from 3D Indoor Reconstructions [94.17683799712397]
We focus on scene graphs, a data structure that organizes the entities of a scene in a graph.
We propose a learned method that regresses a scene graph from the point cloud of a scene.
We show the application of our method in a domain-agnostic retrieval task, where graphs serve as an intermediate representation for 3D-3D and 2D-3D matching.
arXiv Detail & Related papers (2020-04-08T12:25:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.