Inverting the Imaging Process by Learning an Implicit Camera Model
- URL: http://arxiv.org/abs/2304.12748v1
- Date: Tue, 25 Apr 2023 11:55:03 GMT
- Title: Inverting the Imaging Process by Learning an Implicit Camera Model
- Authors: Xin Huang, Qi Zhang, Ying Feng, Hongdong Li, Qing Wang
- Abstract summary: This paper proposes a novel implicit camera model which represents the physical imaging process of a camera as a deep neural network.
We demonstrate the power of this new implicit camera model on two inverse imaging tasks.
- Score: 73.81635386829846
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Representing visual signals with implicit coordinate-based neural networks,
as an effective replacement of the traditional discrete signal representation,
has gained considerable popularity in computer vision and graphics. In contrast
to existing implicit neural representations which focus on modelling the scene
only, this paper proposes a novel implicit camera model which represents the
physical imaging process of a camera as a deep neural network. We demonstrate
the power of this new implicit camera model on two inverse imaging tasks: i)
generating all-in-focus photos, and ii) HDR imaging. Specifically, we devise an
implicit blur generator and an implicit tone mapper to model the aperture and
exposure of the camera's imaging process, respectively. Our implicit camera
model is jointly learned together with implicit scene models under multi-focus
stack and multi-exposure bracket supervision. We have demonstrated the
effectiveness of our new model on a large number of test images and videos,
producing accurate and visually appealing all-in-focus and high dynamic range
images. In principle, our new implicit neural camera model has the potential to
benefit a wide array of other inverse imaging tasks.
Related papers
- Learning Robust Multi-Scale Representation for Neural Radiance Fields
from Unposed Images [65.41966114373373]
We present an improved solution to the neural image-based rendering problem in computer vision.
The proposed approach could synthesize a realistic image of the scene from a novel viewpoint at test time.
arXiv Detail & Related papers (2023-11-08T08:18:23Z) - Robustifying the Multi-Scale Representation of Neural Radiance Fields [86.69338893753886]
We present a robust multi-scale neural radiance fields representation approach to overcome both real-world imaging issues.
Our method handles multi-scale imaging effects and camera-pose estimation problems with NeRF-inspired approaches.
We demonstrate, with examples, that for an accurate neural representation of an object from day-to-day acquired multi-view images, it is crucial to have precise camera-pose estimates.
arXiv Detail & Related papers (2022-10-09T11:46:45Z) - im2nerf: Image to Neural Radiance Field in the Wild [47.18702901448768]
im2nerf is a learning framework that predicts a continuous neural object representation given a single input image in the wild.
We show that im2nerf achieves the state-of-the-art performance for novel view synthesis from a single-view unposed image in the wild.
arXiv Detail & Related papers (2022-09-08T23:28:56Z) - Representing Camera Response Function by a Single Latent Variable and
Fully Connected Neural Network [14.27259159089287]
Modelling the mapping from scene irradiance to image intensity is essential for many computer vision tasks.
New high-performance camera response model that uses a single latent variable and fully connected neural network is proposed.
arXiv Detail & Related papers (2022-09-08T08:02:57Z) - RelPose: Predicting Probabilistic Relative Rotation for Single Objects
in the Wild [73.1276968007689]
We describe a data-driven method for inferring the camera viewpoints given multiple images of an arbitrary object.
We show that our approach outperforms state-of-the-art SfM and SLAM methods given sparse images on both seen and unseen categories.
arXiv Detail & Related papers (2022-08-11T17:59:59Z) - VMRF: View Matching Neural Radiance Fields [57.93631771072756]
VMRF is an innovative view matching NeRF that enables effective NeRF training without requiring prior knowledge in camera poses or camera pose distributions.
VMRF introduces a view matching scheme, which exploits unbalanced optimal transport to produce a feature transport plan for mapping a rendered image with randomly camera pose to the corresponding real image.
With the feature transport plan as the guidance, a novel pose calibration technique is designed which rectifies the initially randomized camera poses by predicting relative pose between the pair of rendered and real images.
arXiv Detail & Related papers (2022-07-06T12:26:40Z) - Single-Image HDR Reconstruction by Learning to Reverse the Camera
Pipeline [100.5353614588565]
We propose to incorporate the domain knowledge of the LDR image formation pipeline into our model.
We model the HDRto-LDR image formation pipeline as the (1) dynamic range clipping, (2) non-linear mapping from a camera response function, and (3) quantization.
We demonstrate that the proposed method performs favorably against state-of-the-art single-image HDR reconstruction algorithms.
arXiv Detail & Related papers (2020-04-02T17:59:04Z) - Text-to-Image Generation with Attention Based Recurrent Neural Networks [1.2599533416395765]
We develop a tractable and stable caption-based image generation model.
Experimentations are performed on Microsoft datasets.
Results show that the proposed model performs better than contemporary approaches.
arXiv Detail & Related papers (2020-01-18T12:19:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.