Remote Sensing Novel View Synthesis with Implicit Multiplane
Representations
- URL: http://arxiv.org/abs/2205.08908v1
- Date: Wed, 18 May 2022 13:03:55 GMT
- Title: Remote Sensing Novel View Synthesis with Implicit Multiplane
Representations
- Authors: Yongchang Wu, Zhengxia Zou, Zhenwei Shi
- Abstract summary: We propose a novel remote sensing view synthesis method by leveraging the recent advances in implicit neural representations.
Considering the overhead and far depth imaging of remote sensing images, we represent the 3D space by combining implicit multiplane images (MPI) representation and deep neural networks.
Images from any novel views can be freely rendered on the basis of the reconstructed model.
- Score: 26.33490094119609
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Novel view synthesis of remote sensing scenes is of great significance for
scene visualization, human-computer interaction, and various downstream
applications. Despite the recent advances in computer graphics and
photogrammetry technology, generating novel views is still challenging
particularly for remote sensing images due to its high complexity, view
sparsity and limited view-perspective variations. In this paper, we propose a
novel remote sensing view synthesis method by leveraging the recent advances in
implicit neural representations. Considering the overhead and far depth imaging
of remote sensing images, we represent the 3D space by combining implicit
multiplane images (MPI) representation and deep neural networks. The 3D scene
is reconstructed under a self-supervised optimization paradigm through a
differentiable multiplane renderer with multi-view input constraints. Images
from any novel views thus can be freely rendered on the basis of the
reconstructed model. As a by-product, the depth maps corresponding to the given
viewpoint can be generated along with the rendering output. We refer to our
method as Implicit Multiplane Images (ImMPI). To further improve the view
synthesis under sparse-view inputs, we explore the learning-based
initialization of remote sensing 3D scenes and proposed a neural network based
Prior extractor to accelerate the optimization process. In addition, we propose
a new dataset for remote sensing novel view synthesis with multi-view
real-world google earth images. Extensive experiments demonstrate the
superiority of the ImMPI over previous state-of-the-art methods in terms of
reconstruction accuracy, visual fidelity, and time efficiency. Ablation
experiments also suggest the effectiveness of our methodology design. Our
dataset and code can be found at https://github.com/wyc-Chang/ImMPI
Related papers
- Multi-Plane Neural Radiance Fields for Novel View Synthesis [5.478764356647437]
Novel view synthesis is a long-standing problem that revolves around rendering frames of scenes from novel camera viewpoints.
In this work, we examine the performance, generalization, and efficiency of single-view multi-plane neural radiance fields.
We propose a new multiplane NeRF architecture that accepts multiple views to improve the synthesis results and expand the viewing range.
arXiv Detail & Related papers (2023-03-03T06:32:55Z) - DINER: Depth-aware Image-based NEural Radiance fields [45.63488428831042]
We present Depth-aware Image-based NEural Radiance fields (DINER)
Given a sparse set of RGB input views, we predict depth and feature maps to guide the reconstruction of a scene representation.
We propose novel techniques to incorporate depth information into feature fusion and efficient scene sampling.
arXiv Detail & Related papers (2022-11-29T23:22:44Z) - High-fidelity 3D GAN Inversion by Pseudo-multi-view Optimization [51.878078860524795]
We present a high-fidelity 3D generative adversarial network (GAN) inversion framework that can synthesize photo-realistic novel views.
Our approach enables high-fidelity 3D rendering from a single image, which is promising for various applications of AI-generated 3D content.
arXiv Detail & Related papers (2022-11-28T18:59:52Z) - Vision Transformer for NeRF-Based View Synthesis from a Single Input
Image [49.956005709863355]
We propose to leverage both the global and local features to form an expressive 3D representation.
To synthesize a novel view, we train a multilayer perceptron (MLP) network conditioned on the learned 3D representation to perform volume rendering.
Our method can render novel views from only a single input image and generalize across multiple object categories using a single model.
arXiv Detail & Related papers (2022-07-12T17:52:04Z) - Scene Representation Transformer: Geometry-Free Novel View Synthesis
Through Set-Latent Scene Representations [48.05445941939446]
A classical problem in computer vision is to infer a 3D scene representation from few images that can be used to render novel views at interactive rates.
We propose the Scene Representation Transformer (SRT), a method which processes posed or unposed RGB images of a new area.
We show that this method outperforms recent baselines in terms of PSNR and speed on synthetic datasets.
arXiv Detail & Related papers (2021-11-25T16:18:56Z) - MVSNeRF: Fast Generalizable Radiance Field Reconstruction from
Multi-View Stereo [52.329580781898116]
We present MVSNeRF, a novel neural rendering approach that can efficiently reconstruct neural radiance fields for view synthesis.
Unlike prior works on neural radiance fields that consider per-scene optimization on densely captured images, we propose a generic deep neural network that can reconstruct radiance fields from only three nearby input views via fast network inference.
arXiv Detail & Related papers (2021-03-29T13:15:23Z) - IBRNet: Learning Multi-View Image-Based Rendering [67.15887251196894]
We present a method that synthesizes novel views of complex scenes by interpolating a sparse set of nearby views.
By drawing on source views at render time, our method hearkens back to classic work on image-based rendering.
arXiv Detail & Related papers (2021-02-25T18:56:21Z) - Neural Body: Implicit Neural Representations with Structured Latent
Codes for Novel View Synthesis of Dynamic Humans [56.63912568777483]
This paper addresses the challenge of novel view synthesis for a human performer from a very sparse set of camera views.
We propose Neural Body, a new human body representation which assumes that the learned neural representations at different frames share the same set of latent codes anchored to a deformable mesh.
Experiments on ZJU-MoCap show that our approach outperforms prior works by a large margin in terms of novel view synthesis quality.
arXiv Detail & Related papers (2020-12-31T18:55:38Z) - NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis [78.5281048849446]
We present a method that achieves state-of-the-art results for synthesizing novel views of complex scenes.
Our algorithm represents a scene using a fully-connected (non-convolutional) deep network.
Because volume rendering is naturally differentiable, the only input required to optimize our representation is a set of images with known camera poses.
arXiv Detail & Related papers (2020-03-19T17:57:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.