SLIDE: Single Image 3D Photography with Soft Layering and Depth-aware
Inpainting
- URL: http://arxiv.org/abs/2109.01068v1
- Date: Thu, 2 Sep 2021 16:37:20 GMT
- Title: SLIDE: Single Image 3D Photography with Soft Layering and Depth-aware
Inpainting
- Authors: Varun Jampani, Huiwen Chang, Kyle Sargent, Abhishek Kar, Richard
Tucker, Michael Krainin, Dominik Kaeser, William T. Freeman, David Salesin,
Brian Curless, Ce Liu
- Abstract summary: Single image 3D photography enables viewers to view a still image from novel viewpoints.
Recent approaches combine monocular depth networks with inpainting networks to achieve compelling results.
We present SLIDE, a modular and unified system for single image 3D photography.
- Score: 54.419266357283966
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Single image 3D photography enables viewers to view a still image from novel
viewpoints. Recent approaches combine monocular depth networks with inpainting
networks to achieve compelling results. A drawback of these techniques is the
use of hard depth layering, making them unable to model intricate appearance
details such as thin hair-like structures. We present SLIDE, a modular and
unified system for single image 3D photography that uses a simple yet effective
soft layering strategy to better preserve appearance details in novel views. In
addition, we propose a novel depth-aware training strategy for our inpainting
module, better suited for the 3D photography task. The resulting SLIDE approach
is modular, enabling the use of other components such as segmentation and
matting for improved layering. At the same time, SLIDE uses an efficient
layered depth formulation that only requires a single forward pass through the
component networks to produce high quality 3D photos. Extensive experimental
analysis on three view-synthesis datasets, in combination with user studies on
in-the-wild image collections, demonstrate superior performance of our
technique in comparison to existing strong baselines while being conceptually
much simpler. Project page: https://varunjampani.github.io/slide
Related papers
- TeMO: Towards Text-Driven 3D Stylization for Multi-Object Meshes [67.5351491691866]
We present a novel framework, dubbed TeMO, to parse multi-object 3D scenes and edit their styles.
Our method can synthesize high-quality stylized content and outperform the existing methods over a wide range of multi-object 3D meshes.
arXiv Detail & Related papers (2023-12-07T12:10:05Z) - Holistic Inverse Rendering of Complex Facade via Aerial 3D Scanning [38.72679977945778]
We use multi-view aerial images to reconstruct the geometry, lighting, and material of facades using neural signed distance fields (SDFs)
The experiment demonstrates the superior quality of our method on facade holistic inverse rendering, novel view synthesis, and scene editing compared to state-of-the-art baselines.
arXiv Detail & Related papers (2023-11-20T15:03:56Z) - TMO: Textured Mesh Acquisition of Objects with a Mobile Device by using
Differentiable Rendering [54.35405028643051]
We present a new pipeline for acquiring a textured mesh in the wild with a single smartphone.
Our method first introduces an RGBD-aided structure from motion, which can yield filtered depth maps.
We adopt the neural implicit surface reconstruction method, which allows for high-quality mesh.
arXiv Detail & Related papers (2023-03-27T10:07:52Z) - High-fidelity 3D GAN Inversion by Pseudo-multi-view Optimization [51.878078860524795]
We present a high-fidelity 3D generative adversarial network (GAN) inversion framework that can synthesize photo-realistic novel views.
Our approach enables high-fidelity 3D rendering from a single image, which is promising for various applications of AI-generated 3D content.
arXiv Detail & Related papers (2022-11-28T18:59:52Z) - Single-View View Synthesis in the Wild with Learned Adaptive Multiplane
Images [15.614631883233898]
Existing methods have shown promising results leveraging monocular depth estimation and color inpainting with layered depth representations.
We propose a new method based on the multiplane image (MPI) representation.
The experiments on both synthetic and real datasets demonstrate that our trained model works remarkably well and achieves state-of-the-art results.
arXiv Detail & Related papers (2022-05-24T02:57:16Z) - Improved Modeling of 3D Shapes with Multi-view Depth Maps [48.8309897766904]
We present a general-purpose framework for modeling 3D shapes using CNNs.
Using just a single depth image of the object, we can output a dense multi-view depth map representation of 3D objects.
arXiv Detail & Related papers (2020-09-07T17:58:27Z) - 3D Photography using Context-aware Layered Depth Inpainting [50.66235795163143]
We propose a method for converting a single RGB-D input image into a 3D photo.
A learning-based inpainting model synthesizes new local color-and-depth content into the occluded region.
The resulting 3D photos can be efficiently rendered with motion parallax.
arXiv Detail & Related papers (2020-04-09T17:59:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.