Dynamic Scene Novel View Synthesis via Deferred Spatio-temporal
Consistency
- URL: http://arxiv.org/abs/2109.01018v1
- Date: Thu, 2 Sep 2021 15:29:45 GMT
- Title: Dynamic Scene Novel View Synthesis via Deferred Spatio-temporal
Consistency
- Authors: Beatrix-Em\H{o}ke F\"ul\"op-Balogh and Eleanor Tursman and James
Tompkin and Julie Digne and Nicolas Bonneel
- Abstract summary: Structures (SfM) and novel view synthesis (NVS) are presented.
SfM produces noisy-temporally reconstructed sparse clouds, resulting in NVS with temporally inconsistent effects.
We demonstrate our algorithm on real-world dynamic scenes against classic more recent learning-based baseline approaches.
- Score: 18.036582072609882
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Structure from motion (SfM) enables us to reconstruct a scene via casual
capture from cameras at different viewpoints, and novel view synthesis (NVS)
allows us to render a captured scene from a new viewpoint. Both are hard with
casual capture and dynamic scenes: SfM produces noisy and spatio-temporally
sparse reconstructed point clouds, resulting in NVS with spatio-temporally
inconsistent effects. We consider SfM and NVS parts together to ease the
challenge. First, for SfM, we recover stable camera poses, then we defer the
requirement for temporally-consistent points across the scene and reconstruct
only a sparse point cloud per timestep that is noisy in space-time. Second, for
NVS, we present a variational diffusion formulation on depths and colors that
lets us robustly cope with the noise by enforcing spatio-temporal consistency
via per-pixel reprojection weights derived from the input views. Together, this
deferred approach generates novel views for dynamic scenes without requiring
challenging spatio-temporally consistent reconstructions nor training complex
models on large datasets. We demonstrate our algorithm on real-world dynamic
scenes against classic and more recent learning-based baseline approaches.
Related papers
- D-NPC: Dynamic Neural Point Clouds for Non-Rigid View Synthesis from Monocular Video [53.83936023443193]
This paper contributes to the field by introducing a new synthesis method for dynamic novel view from monocular video, such as smartphone captures.
Our approach represents the as a $textitdynamic neural point cloud$, an implicit time-conditioned point cloud that encodes local geometry and appearance in separate hash-encoded neural feature grids.
arXiv Detail & Related papers (2024-06-14T14:35:44Z) - SceNeRFlow: Time-Consistent Reconstruction of General Dynamic Scenes [75.9110646062442]
We propose SceNeRFlow to reconstruct a general, non-rigid scene in a time-consistent manner.
Our method takes multi-view RGB videos and background images from static cameras with known camera parameters as input.
We show experimentally that, unlike prior work that only handles small motion, our method enables the reconstruction of studio-scale motions.
arXiv Detail & Related papers (2023-08-16T09:50:35Z) - DynIBaR: Neural Dynamic Image-Based Rendering [79.44655794967741]
We address the problem of synthesizing novel views from a monocular video depicting a complex dynamic scene.
We adopt a volumetric image-based rendering framework that synthesizes new viewpoints by aggregating features from nearby views.
We demonstrate significant improvements over state-of-the-art methods on dynamic scene datasets.
arXiv Detail & Related papers (2022-11-20T20:57:02Z) - Learning Dynamic View Synthesis With Few RGBD Cameras [60.36357774688289]
We propose to utilize RGBD cameras to synthesize free-viewpoint videos of dynamic indoor scenes.
We generate point clouds from RGBD frames and then render them into free-viewpoint videos via a neural feature.
We introduce a simple Regional Depth-Inpainting module that adaptively inpaints missing depth values to render complete novel views.
arXiv Detail & Related papers (2022-04-22T03:17:35Z) - Dynamic View Synthesis from Dynamic Monocular Video [69.80425724448344]
We present an algorithm for generating views at arbitrary viewpoints and any input time step given a monocular video of a dynamic scene.
We show extensive quantitative and qualitative results of dynamic view synthesis from casually captured videos.
arXiv Detail & Related papers (2021-05-13T17:59:50Z) - Neural Trajectory Fields for Dynamic Novel View Synthesis [40.9665251865609]
We introduce DCT-NeRF, a coordinatebased neural representation for dynamic scenes.
We learn smooth and stable trajectories over the input sequence for each point in space.
This allows us to enforce consistency between any two frames in the sequence, which results in high quality reconstruction.
arXiv Detail & Related papers (2021-05-12T22:38:30Z) - Non-Rigid Neural Radiance Fields: Reconstruction and Novel View
Synthesis of a Dynamic Scene From Monocular Video [76.19076002661157]
Non-Rigid Neural Radiance Fields (NR-NeRF) is a reconstruction and novel view synthesis approach for general non-rigid dynamic scenes.
We show that even a single consumer-grade camera is sufficient to synthesize sophisticated renderings of a dynamic scene from novel virtual camera views.
arXiv Detail & Related papers (2020-12-22T18:46:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.