Adaptive Multiplane Image Generation from a Single Internet Picture
- URL: http://arxiv.org/abs/2011.13317v1
- Date: Thu, 26 Nov 2020 14:35:05 GMT
- Title: Adaptive Multiplane Image Generation from a Single Internet Picture
- Authors: Diogo C. Luvizon, Gustavo Sutter P. Carvalho, Andreza A. dos Santos,
Jhonatas S. Conceicao, Jose L. Flores-Campana, Luis G. L. Decker, Marcos R.
Souza, Helio Pedrini, Antonio Joia, Otavio A. B. Penatti
- Abstract summary: We address the problem of generating a multiplane image (MPI) from a single high-resolution picture.
We propose an adaptive slicing algorithm that produces an MPI with a variable number of image planes.
We show that our method is capable of producing high-quality predictions with one order of magnitude less parameters compared to previous approaches.
- Score: 1.8961324344454253
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the last few years, several works have tackled the problem of novel view
synthesis from stereo images or even from a single picture. However, previous
methods are computationally expensive, specially for high-resolution images. In
this paper, we address the problem of generating a multiplane image (MPI) from
a single high-resolution picture. We present the adaptive-MPI representation,
which allows rendering novel views with low computational requirements. To this
end, we propose an adaptive slicing algorithm that produces an MPI with a
variable number of image planes. We present a new lightweight CNN for depth
estimation, which is learned by knowledge distillation from a larger network.
Occluded regions in the adaptive-MPI are inpainted also by a lightweight CNN.
We show that our method is capable of producing high-quality predictions with
one order of magnitude less parameters compared to previous approaches. The
robustness of our method is evidenced on challenging pictures from the
Internet.
Related papers
- Parameter-Inverted Image Pyramid Networks [49.35689698870247]
We propose a novel network architecture known as the Inverted Image Pyramid Networks (PIIP)
Our core idea is to use models with different parameter sizes to process different resolution levels of the image pyramid.
PIIP achieves superior performance in tasks such as object detection, segmentation, and image classification.
arXiv Detail & Related papers (2024-06-06T17:59:10Z) - Hyper-VolTran: Fast and Generalizable One-Shot Image to 3D Object
Structure via HyperNetworks [53.67497327319569]
We introduce a novel neural rendering technique to solve image-to-3D from a single view.
Our approach employs the signed distance function as the surface representation and incorporates generalizable priors through geometry-encoding volumes and HyperNetworks.
Our experiments show the advantages of our proposed approach with consistent results and rapid generation.
arXiv Detail & Related papers (2023-12-24T08:42:37Z) - MMPI: a Flexible Radiance Field Representation by Multiple Multi-plane
Images Blending [61.45757368117578]
This paper presents a flexible representation of neural radiance fields based on multi-plane images (MPI)
MPI is widely used in NeRF learning for its simple definition, easy calculation, and powerful ability to represent unbounded scenes.
We show that MPI can synthesize high-quality novel views of complex scenes with diverse camera distributions and view directions.
arXiv Detail & Related papers (2023-09-30T04:36:43Z) - T-former: An Efficient Transformer for Image Inpainting [50.43302925662507]
A class of attention-based network architectures, called transformer, has shown significant performance on natural language processing fields.
In this paper, we design a novel attention linearly related to the resolution according to Taylor expansion, and based on this attention, a network called $T$-former is designed for image inpainting.
Experiments on several benchmark datasets demonstrate that our proposed method achieves state-of-the-art accuracy while maintaining a relatively low number of parameters and computational complexity.
arXiv Detail & Related papers (2023-05-12T04:10:42Z) - Efficient Textured Mesh Recovery from Multiple Views with Differentiable
Rendering [8.264851594332677]
We propose an efficient coarse-to-fine approach to recover the textured mesh from multi-view images.
We optimize the shape geometry by minimizing the difference between the rendered mesh with the depth predicted by the learning-based multi-view stereo algorithm.
In contrast to the implicit neural representation on shape and color, we introduce a physically based inverse rendering scheme to jointly estimate the lighting and reflectance of the objects.
arXiv Detail & Related papers (2022-05-25T03:33:55Z) - An Optical physics inspired CNN approach for intrinsic image
decomposition [0.0]
Intrinsic Image Decomposition is an open problem of generating the constituents of an image.
We propose a neural network architecture capable of this decomposition using physics-based parameters derived from the image.
arXiv Detail & Related papers (2021-05-21T00:54:01Z) - Deep Variational Network Toward Blind Image Restoration [60.45350399661175]
Blind image restoration is a common yet challenging problem in computer vision.
We propose a novel blind image restoration method, aiming to integrate both the advantages of them.
Experiments on two typical blind IR tasks, namely image denoising and super-resolution, demonstrate that the proposed method achieves superior performance over current state-of-the-arts.
arXiv Detail & Related papers (2020-08-25T03:30:53Z) - The Power of Triply Complementary Priors for Image Compressive Sensing [89.14144796591685]
We propose a joint low-rank deep (LRD) image model, which contains a pair of complementaryly trip priors.
We then propose a novel hybrid plug-and-play framework based on the LRD model for image CS.
To make the optimization tractable, a simple yet effective algorithm is proposed to solve the proposed H-based image CS problem.
arXiv Detail & Related papers (2020-05-16T08:17:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.