Video Reconstruction by Spatio-Temporal Fusion of Blurred-Coded Image
Pair
- URL: http://arxiv.org/abs/2010.10052v2
- Date: Fri, 13 Nov 2020 10:06:06 GMT
- Title: Video Reconstruction by Spatio-Temporal Fusion of Blurred-Coded Image
Pair
- Authors: S Anupama, Prasan Shedligeri, Abhishek Pal, Kaushik Mitra
- Abstract summary: Recovering video from a single motion-blurred image is a very ill-posed problem.
Traditional coded exposure framework is better-posed but it only samples a fraction of the space-time volume.
We propose to use the complementary information present in the fully-exposed image along with the coded exposure image to recover a high fidelity video.
- Score: 16.295479896947853
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Learning-based methods have enabled the recovery of a video sequence from a
single motion-blurred image or a single coded exposure image. Recovering video
from a single motion-blurred image is a very ill-posed problem and the
recovered video usually has many artifacts. In addition to this, the direction
of motion is lost and it results in motion ambiguity. However, it has the
advantage of fully preserving the information in the static parts of the scene.
The traditional coded exposure framework is better-posed but it only samples a
fraction of the space-time volume, which is at best 50% of the space-time
volume. Here, we propose to use the complementary information present in the
fully-exposed (blurred) image along with the coded exposure image to recover a
high fidelity video without any motion ambiguity. Our framework consists of a
shared encoder followed by an attention module to selectively combine the
spatial information from the fully-exposed image with the temporal information
from the coded image, which is then super-resolved to recover a non-ambiguous
high-quality video. The input to our algorithm is a fully-exposed and coded
image pair. Such an acquisition system already exists in the form of a
Coded-two-bucket (C2B) camera. We demonstrate that our proposed deep learning
approach using blurred-coded image pair produces much better results than those
from just a blurred image or just a coded image.
Related papers
- Buffer Anytime: Zero-Shot Video Depth and Normal from Image Priors [54.8852848659663]
Buffer Anytime is a framework for estimation of depth and normal maps (which we call geometric buffers) from video.
We demonstrate high-quality video buffer estimation by leveraging single-image priors with temporal consistency constraints.
arXiv Detail & Related papers (2024-11-26T09:28:32Z) - SITAR: Semi-supervised Image Transformer for Action Recognition [20.609596080624662]
This paper addresses video action recognition in a semi-supervised setting by leveraging only a handful of labeled videos.
We capitalize on the vast pool of unlabeled samples and employ contrastive learning on the encoded super images.
Our method demonstrates superior performance compared to existing state-of-the-art approaches for semi-supervised action recognition.
arXiv Detail & Related papers (2024-09-04T17:49:54Z) - Neuromorphic Synergy for Video Binarization [54.195375576583864]
Bimodal objects serve as a visual form to embed information that can be easily recognized by vision systems.
Neuromorphic cameras offer new capabilities for alleviating motion blur, but it is non-trivial to first de-blur and then binarize the images in a real-time manner.
We propose an event-based binary reconstruction method that leverages the prior knowledge of the bimodal target's properties to perform inference independently in both event space and image space.
We also develop an efficient integration method to propagate this binary image to high frame rate binary video.
arXiv Detail & Related papers (2024-02-20T01:43:51Z) - Lightweight High-Speed Photography Built on Coded Exposure and Implicit Neural Representation of Videos [34.152901518593396]
The demand for compact cameras capable of recording high-speed scenes with high resolution is steadily increasing.
However, achieving such capabilities often entails high bandwidth requirements, resulting in bulky, heavy systems unsuitable for low-capacity platforms.
We propose a novel approach to address these challenges by combining the classical coded exposure imaging technique with the emerging implicit neural representation for videos.
arXiv Detail & Related papers (2023-11-22T03:41:13Z) - Neural Image Re-Exposure [86.42475408644822]
An improper shutter may lead to a blurry image, video discontinuity, or rolling shutter artifact.
We propose a neural network-based image re-exposure framework.
It consists of an encoder for visual latent space construction, a re-exposure module for aggregating information to neural film with a desired shutter strategy, and a decoder for 'developing' neural film into a desired image.
arXiv Detail & Related papers (2023-05-23T01:55:37Z) - Joint Video Multi-Frame Interpolation and Deblurring under Unknown
Exposure Time [101.91824315554682]
In this work, we aim ambitiously for a more realistic and challenging task - joint video multi-frame and deblurring under unknown exposure time.
We first adopt a variant of supervised contrastive learning to construct an exposure-aware representation from input blurred frames.
We then build our video reconstruction network upon the exposure and motion representation by progressive exposure-adaptive convolution and motion refinement.
arXiv Detail & Related papers (2023-03-27T09:43:42Z) - Unfolding a blurred image [36.519356428362286]
We learn motion representation from sharp videos in an unsupervised manner.
We then train a convolutional recurrent video autoencoder network that performs a surrogate task of video reconstruction.
It is employed for guided training of a motion encoder for blurred images.
This network extracts embedded motion information from the blurred image to generate a sharp video in conjunction with the trained recurrent video decoder.
arXiv Detail & Related papers (2022-01-28T09:39:55Z) - Restoration of Video Frames from a Single Blurred Image with Motion
Understanding [69.90724075337194]
We propose a novel framework to generate clean video frames from a single motion-red image.
We formulate video restoration from a single blurred image as an inverse problem by setting clean image sequence and their respective motion as latent factors.
Our framework is based on anblur-decoder structure with spatial transformer network modules.
arXiv Detail & Related papers (2021-04-19T08:32:57Z) - Motion-blurred Video Interpolation and Extrapolation [72.3254384191509]
We present a novel framework for deblurring, interpolating and extrapolating sharp frames from a motion-blurred video in an end-to-end manner.
To ensure temporal coherence across predicted frames and address potential temporal ambiguity, we propose a simple, yet effective flow-based rule.
arXiv Detail & Related papers (2021-03-04T12:18:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.