SinGRAF: Learning a 3D Generative Radiance Field for a Single Scene
- URL: http://arxiv.org/abs/2211.17260v2
- Date: Sun, 2 Apr 2023 14:26:57 GMT
- Title: SinGRAF: Learning a 3D Generative Radiance Field for a Single Scene
- Authors: Minjung Son, Jeong Joon Park, Leonidas Guibas, Gordon Wetzstein
- Abstract summary: We introduce SinGRAF, a 3D-aware generative model that is trained with a few input images of a single scene.
It generates different realizations of this 3D scene that preserve the appearance of the input while varying scene layout.
With several experiments, we demonstrate that the results produced by SinGRAF outperform the closest related works in both quality and diversity by a large margin.
- Score: 40.705096946588
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Generative models have shown great promise in synthesizing photorealistic 3D
objects, but they require large amounts of training data. We introduce SinGRAF,
a 3D-aware generative model that is trained with a few input images of a single
scene. Once trained, SinGRAF generates different realizations of this 3D scene
that preserve the appearance of the input while varying scene layout. For this
purpose, we build on recent progress in 3D GAN architectures and introduce a
novel progressive-scale patch discrimination approach during training. With
several experiments, we demonstrate that the results produced by SinGRAF
outperform the closest related works in both quality and diversity by a large
margin.
Related papers
- Visibility-Uncertainty-guided 3D Gaussian Inpainting via Scene Conceptional Learning [63.94919846010485]
3D Gaussian inpainting (3DGI) is challenging in effectively leveraging complementary visual and semantic cues from multiple input views.
We propose a method that measures the visibility uncertainties of 3D points across different input views and uses them to guide 3DGI.
We build a novel 3DGI framework, VISTA, by integrating VISibility-uncerTainty-guided 3DGI with scene conceptuAl learning.
arXiv Detail & Related papers (2025-04-23T06:21:11Z) - A Recipe for Generating 3D Worlds From a Single Image [28.396381735501524]
We introduce a recipe for generating immersive 3D worlds from a single image.
This approach requires minimal training and uses existing generative models.
Tested on both synthetic and real images, our method produces high-quality 3D environments suitable for VR display.
arXiv Detail & Related papers (2025-03-20T18:06:12Z) - Invisible Stitch: Generating Smooth 3D Scenes with Depth Inpainting [75.7154104065613]
We introduce a novel depth completion model, trained via teacher distillation and self-training to learn the 3D fusion process.
We also introduce a new benchmarking scheme for scene generation methods that is based on ground truth geometry.
arXiv Detail & Related papers (2024-04-30T17:59:40Z) - Bootstrap 3D Reconstructed Scenes from 3D Gaussian Splatting [10.06208115191838]
We present a bootstrapping method to enhance the rendering of novel views using trained 3D-GS.
Our results indicate that bootstrapping effectively reduces artifacts, as well as clear enhancements on the evaluation metrics.
arXiv Detail & Related papers (2024-04-29T12:57:05Z) - 3D-SceneDreamer: Text-Driven 3D-Consistent Scene Generation [51.64796781728106]
We propose a generative refinement network to synthesize new contents with higher quality by exploiting the natural image prior to 2D diffusion model and the global 3D information of the current scene.
Our approach supports wide variety of scene generation and arbitrary camera trajectories with improved visual quality and 3D consistency.
arXiv Detail & Related papers (2024-03-14T14:31:22Z) - Denoising Diffusion via Image-Based Rendering [54.20828696348574]
We introduce the first diffusion model able to perform fast, detailed reconstruction and generation of real-world 3D scenes.
First, we introduce a new neural scene representation, IB-planes, that can efficiently and accurately represent large 3D scenes.
Second, we propose a denoising-diffusion framework to learn a prior over this novel 3D scene representation, using only 2D images.
arXiv Detail & Related papers (2024-02-05T19:00:45Z) - PonderV2: Pave the Way for 3D Foundation Model with A Universal
Pre-training Paradigm [114.47216525866435]
We introduce a novel universal 3D pre-training framework designed to facilitate the acquisition of efficient 3D representation.
For the first time, PonderV2 achieves state-of-the-art performance on 11 indoor and outdoor benchmarks, implying its effectiveness.
arXiv Detail & Related papers (2023-10-12T17:59:57Z) - 3inGAN: Learning a 3D Generative Model from Images of a Self-similar
Scene [34.2144933185175]
3inGAN is an unconditional 3D generative model trained from 2D images of a single self-similar 3D scene.
We show results on semi-stochastic scenes of varying scale and complexity, obtained from real and synthetic sources.
arXiv Detail & Related papers (2022-11-27T18:03:21Z) - 3D-Aware Indoor Scene Synthesis with Depth Priors [62.82867334012399]
Existing methods fail to model indoor scenes due to the large diversity of room layouts and the objects inside.
We argue that indoor scenes do not have a shared intrinsic structure, and hence only using 2D images cannot adequately guide the model with the 3D geometry.
arXiv Detail & Related papers (2022-02-17T09:54:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.