Infinite Nature: Perpetual View Generation of Natural Scenes from a
Single Image
- URL: http://arxiv.org/abs/2012.09855v2
- Date: Fri, 18 Dec 2020 05:49:19 GMT
- Title: Infinite Nature: Perpetual View Generation of Natural Scenes from a
Single Image
- Authors: Andrew Liu, Richard Tucker, Varun Jampani, Ameesh Makadia, Noah
Snavely, Angjoo Kanazawa
- Abstract summary: We introduce the problem of perpetual view generation -- long-range generation of novel views corresponding to an arbitrarily long camera trajectory given a single image.
We take a hybrid approach that integrates both geometry and image synthesis in an iterative render, refine, and repeat framework.
Our approach can be trained from a set of monocular video sequences without any manual annotation.
- Score: 73.56631858393148
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: We introduce the problem of perpetual view generation -- long-range
generation of novel views corresponding to an arbitrarily long camera
trajectory given a single image. This is a challenging problem that goes far
beyond the capabilities of current view synthesis methods, which work for a
limited range of viewpoints and quickly degenerate when presented with a large
camera motion. Methods designed for video generation also have limited ability
to produce long video sequences and are often agnostic to scene geometry. We
take a hybrid approach that integrates both geometry and image synthesis in an
iterative render, refine, and repeat framework, allowing for long-range
generation that cover large distances after hundreds of frames. Our approach
can be trained from a set of monocular video sequences without any manual
annotation. We propose a dataset of aerial footage of natural coastal scenes,
and compare our method with recent view synthesis and conditional video
generation baselines, showing that it can generate plausible scenes for much
longer time horizons over large camera trajectories compared to existing
methods. Please visit our project page at https://infinite-nature.github.io/.
Related papers
- Cavia: Camera-controllable Multi-view Video Diffusion with View-Integrated Attention [62.2447324481159]
Cavia is a novel framework for camera-controllable, multi-view video generation.
Our framework extends the spatial and temporal attention modules, improving both viewpoint and temporal consistency.
Cavia is the first of its kind that allows the user to specify distinct camera motion while obtaining object motion.
arXiv Detail & Related papers (2024-10-14T17:46:32Z) - Streetscapes: Large-scale Consistent Street View Generation Using Autoregressive Video Diffusion [61.929653153389964]
We present a method for generating Streetscapes-long sequences of views through an on-the-fly synthesized city-scale scene.
Our method can scale to much longer-range camera trajectories, spanning several city blocks, while maintaining visual quality and consistency.
arXiv Detail & Related papers (2024-07-18T17:56:30Z) - Explorative Inbetweening of Time and Space [46.77750028273578]
We introduce bounded generation to control video generation based only on a given start and end frame.
Time Reversal Fusion fuses the temporally forward and backward denoising paths conditioned on the start and end frame.
We find that Time Reversal Fusion outperforms related work on all subtasks.
arXiv Detail & Related papers (2024-03-21T17:57:31Z) - Fast View Synthesis of Casual Videos with Soup-of-Planes [24.35962788109883]
Novel view synthesis from an in-the-wild video is difficult due to challenges like scene dynamics and lack of parallax.
This paper revisits explicit video representations to synthesize high-quality novel views from a monocular video efficiently.
Our method can render high-quality novel views from an in-the-wild video with comparable quality to state-of-the-art methods while being 100x faster in training and enabling real-time rendering.
arXiv Detail & Related papers (2023-12-04T18:55:48Z) - Multi-object Video Generation from Single Frame Layouts [84.55806837855846]
We propose a video generative framework capable of synthesizing global scenes with local objects.
Our framework is a non-trivial adaptation from image generation methods, and is new to this field.
Our model has been evaluated on two widely-used video recognition benchmarks.
arXiv Detail & Related papers (2023-05-06T09:07:01Z) - Towards Smooth Video Composition [59.134911550142455]
Video generation requires consistent and persistent frames with dynamic content over time.
This work investigates modeling the temporal relations for composing video with arbitrary length, from a few frames to even infinite, using generative adversarial networks (GANs)
We show that the alias-free operation for single image generation, together with adequately pre-learned knowledge, brings a smooth frame transition without compromising the per-frame quality.
arXiv Detail & Related papers (2022-12-14T18:54:13Z) - InfiniteNature-Zero: Learning Perpetual View Generation of Natural
Scenes from Single Images [83.37640073416749]
We present a method for learning to generate flythrough videos of natural scenes starting from a single view.
This capability is learned from a collection of single photographs, without requiring camera poses or even multiple views of each scene.
arXiv Detail & Related papers (2022-07-22T15:41:06Z) - Look Outside the Room: Synthesizing A Consistent Long-Term 3D Scene
Video from A Single Image [8.13564646389987]
We propose a novel approach to synthesize a consistent long-term video given a single scene image and a trajectory of large camera motions.
Our method outperforms state-of-the-art view synthesis approaches by a large margin.
arXiv Detail & Related papers (2022-03-17T17:16:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.