Acquiring a Dynamic Light Field through a Single-Shot Coded Image
- URL: http://arxiv.org/abs/2204.12089v1
- Date: Tue, 26 Apr 2022 06:00:02 GMT
- Title: Acquiring a Dynamic Light Field through a Single-Shot Coded Image
- Authors: Ryoya Mizuno, Keita Takahashi, Michitaka Yoshida, Chihiro Tsutake,
Toshiaki Fujii, Hajime Nagahara
- Abstract summary: We propose a method for compressively acquiring a dynamic light field (a 5-D volume) through a single-shot coded image (a 2-D measurement)
We designed an imaging model that synchronously applies aperture coding and pixel-wise exposure coding within a single exposure time.
The observed image is then fed to a convolutional neural network (CNN) for light-field reconstruction, which is jointly trained with the camera-side coding patterns.
- Score: 12.615509935080434
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a method for compressively acquiring a dynamic light field (a 5-D
volume) through a single-shot coded image (a 2-D measurement). We designed an
imaging model that synchronously applies aperture coding and pixel-wise
exposure coding within a single exposure time. This coding scheme enables us to
effectively embed the original information into a single observed image. The
observed image is then fed to a convolutional neural network (CNN) for
light-field reconstruction, which is jointly trained with the camera-side
coding patterns. We also developed a hardware prototype to capture a real 3-D
scene moving over time. We succeeded in acquiring a dynamic light field with
5x5 viewpoints over 4 temporal sub-frames (100 views in total) from a single
observed image. Repeating capture and reconstruction processes over time, we
can acquire a dynamic light field at 4x the frame rate of the camera. To our
knowledge, our method is the first to achieve a finer temporal resolution than
the camera itself in compressive light-field acquisition. Our software is
available from our project webpage
Related papers
- Time-Efficient Light-Field Acquisition Using Coded Aperture and Events [16.130950260664285]
Our method applies a sequence of coding patterns during a single exposure for an image frame.
The parallax information, which is related to the differences in coding patterns, is recorded as events.
The image frame and events, all of which are measured in a single exposure, are jointly used to computationally reconstruct a light field.
arXiv Detail & Related papers (2024-03-12T02:04:17Z) - Reconstructing Continuous Light Field From Single Coded Image [7.937367109582907]
We propose a method for reconstructing a continuous light field of a target scene from a single observed image.
Joint aperture-exposure coding implemented in a camera enables effective embedding of 3-D scene information into an observed image.
NeRF-based neural rendering enables high quality view synthesis of a 3-D scene from continuous viewpoints.
arXiv Detail & Related papers (2023-11-16T07:59:01Z) - Joint 3D Shape and Motion Estimation from Rolling Shutter Light-Field
Images [2.0277446818410994]
We propose an approach to address the problem of 3D reconstruction of scenes from a single image captured by a light-field camera equipped with a rolling shutter sensor.
Our method leverages the 3D information cues present in the light-field and the motion information provided by the rolling shutter effect.
We present a generic model for the imaging process of this sensor and a two-stage algorithm that minimizes the re-projection error.
arXiv Detail & Related papers (2023-11-02T15:08:18Z) - Real-Time Radiance Fields for Single-Image Portrait View Synthesis [85.32826349697972]
We present a one-shot method to infer and render a 3D representation from a single unposed image in real-time.
Given a single RGB input, our image encoder directly predicts a canonical triplane representation of a neural radiance field for 3D-aware novel view synthesis via volume rendering.
Our method is fast (24 fps) on consumer hardware, and produces higher quality results than strong GAN-inversion baselines that require test-time optimization.
arXiv Detail & Related papers (2023-05-03T17:56:01Z) - Joint Video Multi-Frame Interpolation and Deblurring under Unknown
Exposure Time [101.91824315554682]
In this work, we aim ambitiously for a more realistic and challenging task - joint video multi-frame and deblurring under unknown exposure time.
We first adopt a variant of supervised contrastive learning to construct an exposure-aware representation from input blurred frames.
We then build our video reconstruction network upon the exposure and motion representation by progressive exposure-adaptive convolution and motion refinement.
arXiv Detail & Related papers (2023-03-27T09:43:42Z) - Progressively Optimized Local Radiance Fields for Robust View Synthesis [76.55036080270347]
We present an algorithm for reconstructing the radiance field of a large-scale scene from a single casually captured video.
For handling unknown poses, we jointly estimate the camera poses with radiance field in a progressive manner.
For handling large unbounded scenes, we dynamically allocate new local radiance fields trained with frames within a temporal window.
arXiv Detail & Related papers (2023-03-24T04:03:55Z) - NeRFPlayer: A Streamable Dynamic Scene Representation with Decomposed
Neural Radiance Fields [99.57774680640581]
We present an efficient framework capable of fast reconstruction, compact modeling, and streamable rendering.
We propose to decompose the 4D space according to temporal characteristics. Points in the 4D space are associated with probabilities belonging to three categories: static, deforming, and new areas.
arXiv Detail & Related papers (2022-10-28T07:11:05Z) - A Portable Multiscopic Camera for Novel View and Time Synthesis in
Dynamic Scenes [42.00094186447837]
We present a portable multiscopic camera system with a dedicated model for novel view and time synthesis in dynamic scenes.
Our goal is to render high-quality images for a dynamic scene from any viewpoint at any time using our portable multiscopic camera.
arXiv Detail & Related papers (2022-08-30T17:53:17Z) - Real-time dense 3D Reconstruction from monocular video data captured by
low-cost UAVs [0.3867363075280543]
Real-time 3D reconstruction enables fast dense mapping of the environment which benefits numerous applications, such as navigation or live evaluation of an emergency.
In contrast to most real-time capable approaches, our approach does not need an explicit depth sensor.
By exploiting the self-motion of the unmanned aerial vehicle (UAV) flying with oblique view around buildings, we estimate both camera trajectory and depth for selected images with enough novel content.
arXiv Detail & Related papers (2021-04-21T13:12:17Z) - Neural Radiance Flow for 4D View Synthesis and Video Processing [59.9116932930108]
We present a method to learn a 4D spatial-temporal representation of a dynamic scene from a set of RGB images.
Key to our approach is the use of a neural implicit representation that learns to capture the 3D occupancy, radiance, and dynamics of the scene.
arXiv Detail & Related papers (2020-12-17T17:54:32Z) - Lightweight Multi-View 3D Pose Estimation through Camera-Disentangled
Representation [57.11299763566534]
We present a solution to recover 3D pose from multi-view images captured with spatially calibrated cameras.
We exploit 3D geometry to fuse input images into a unified latent representation of pose, which is disentangled from camera view-points.
Our architecture then conditions the learned representation on camera projection operators to produce accurate per-view 2d detections.
arXiv Detail & Related papers (2020-04-05T12:52:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.