Novel View Synthesis of Dynamic Scenes with Globally Coherent Depths
from a Monocular Camera
- URL: http://arxiv.org/abs/2004.01294v1
- Date: Thu, 2 Apr 2020 22:45:53 GMT
- Title: Novel View Synthesis of Dynamic Scenes with Globally Coherent Depths
from a Monocular Camera
- Authors: Jae Shin Yoon, Kihwan Kim, Orazio Gallo, Hyun Soo Park, Jan Kautz
- Abstract summary: This paper presents a new method to synthesize an image from arbitrary views and times given a collection of images of a dynamic scene.
A key challenge for the novel view synthesis arises from dynamic scene reconstruction where epipolar geometry does not apply to the local motion of dynamic contents.
To address this challenge, we propose to combine the depth from single view (DSV) and the depth from multi-view stereo (DMV), where DSV is complete, i.e., a depth is assigned to every pixel, yet view-variant in its scale, while DMV is view-invariant yet incomplete.
- Score: 93.04135520894631
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper presents a new method to synthesize an image from arbitrary views
and times given a collection of images of a dynamic scene. A key challenge for
the novel view synthesis arises from dynamic scene reconstruction where
epipolar geometry does not apply to the local motion of dynamic contents. To
address this challenge, we propose to combine the depth from single view (DSV)
and the depth from multi-view stereo (DMV), where DSV is complete, i.e., a
depth is assigned to every pixel, yet view-variant in its scale, while DMV is
view-invariant yet incomplete. Our insight is that although its scale and
quality are inconsistent with other views, the depth estimation from a single
view can be used to reason about the globally coherent geometry of dynamic
contents. We cast this problem as learning to correct the scale of DSV, and to
refine each depth with locally consistent motions between views to form a
coherent depth estimation. We integrate these tasks into a depth fusion network
in a self-supervised fashion. Given the fused depth maps, we synthesize a
photorealistic virtual view in a specific location and time with our deep
blending network that completes the scene and renders the virtual view. We
evaluate our method of depth estimation and view synthesis on diverse
real-world dynamic scenes and show the outstanding performance over existing
methods.
Related papers
- Refinement of Monocular Depth Maps via Multi-View Differentiable Rendering [4.717325308876748]
We present a novel approach to generate view consistent and detailed depth maps from a number of posed images.
We leverage advances in monocular depth estimation, which generate topologically complete, but metrically inaccurate depth maps.
Our method is able to generate dense, detailed, high-quality depth maps, also in challenging indoor scenarios, and outperforms state-of-the-art depth reconstruction approaches.
arXiv Detail & Related papers (2024-10-04T18:50:28Z) - Dynamic Scene Understanding through Object-Centric Voxelization and Neural Rendering [57.895846642868904]
We present a 3D generative model named DynaVol-S for dynamic scenes that enables object-centric learning.
voxelization infers per-object occupancy probabilities at individual spatial locations.
Our approach integrates 2D semantic features to create 3D semantic grids, representing the scene through multiple disentangled voxel grids.
arXiv Detail & Related papers (2024-07-30T15:33:58Z) - DeLiRa: Self-Supervised Depth, Light, and Radiance Fields [32.350984950639656]
Differentiable volumetric rendering is a powerful paradigm for 3D reconstruction and novel view synthesis.
Standard volume rendering approaches struggle with degenerate geometries in the case of limited viewpoint diversity.
In this work, we propose to use the multi-view photometric objective as a geometric regularizer for volumetric rendering.
arXiv Detail & Related papers (2023-04-06T00:16:25Z) - Decoupling Dynamic Monocular Videos for Dynamic View Synthesis [50.93409250217699]
We tackle the challenge of dynamic view synthesis from dynamic monocular videos in an unsupervised fashion.
Specifically, we decouple the motion of the dynamic objects into object motion and camera motion, respectively regularized by proposed unsupervised surface consistency and patch-based multi-view constraints.
arXiv Detail & Related papers (2023-04-04T11:25:44Z) - Inter-View Depth Consistency Testing in Depth Difference Subspace [6.205922305859478]
Multiview depth imagery will play a critical role in free-viewpoint television.
This paper proposes a method for depth consistency testing in depth difference subspace.
We also propose a view synthesis algorithm that uses the obtained consistency information to improve the visual quality of virtual views.
arXiv Detail & Related papers (2023-01-27T18:43:38Z) - DynIBaR: Neural Dynamic Image-Based Rendering [79.44655794967741]
We address the problem of synthesizing novel views from a monocular video depicting a complex dynamic scene.
We adopt a volumetric image-based rendering framework that synthesizes new viewpoints by aggregating features from nearby views.
We demonstrate significant improvements over state-of-the-art methods on dynamic scene datasets.
arXiv Detail & Related papers (2022-11-20T20:57:02Z) - DnD: Dense Depth Estimation in Crowded Dynamic Indoor Scenes [68.38952377590499]
We present a novel approach for estimating depth from a monocular camera as it moves through complex indoor environments.
Our approach predicts absolute scale depth maps over the entire scene consisting of a static background and multiple moving people.
arXiv Detail & Related papers (2021-08-12T09:12:39Z) - Non-Rigid Neural Radiance Fields: Reconstruction and Novel View
Synthesis of a Dynamic Scene From Monocular Video [76.19076002661157]
Non-Rigid Neural Radiance Fields (NR-NeRF) is a reconstruction and novel view synthesis approach for general non-rigid dynamic scenes.
We show that even a single consumer-grade camera is sufficient to synthesize sophisticated renderings of a dynamic scene from novel virtual camera views.
arXiv Detail & Related papers (2020-12-22T18:46:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.