Low Light Video Enhancement by Learning on Static Videos with
Cross-Frame Attention
- URL: http://arxiv.org/abs/2210.04290v1
- Date: Sun, 9 Oct 2022 15:49:46 GMT
- Title: Low Light Video Enhancement by Learning on Static Videos with
Cross-Frame Attention
- Authors: Shivam Chhirolya, Sameer Malik, Rajiv Soundararajan
- Abstract summary: We develop a deep learning method for low light video enhancement by training a model on static videos.
Existing methods operate frame by frame and do not exploit the relationships among neighbouring frames.
We show that our method outperforms other state-of-the-art video enhancement algorithms when trained only on static videos.
- Score: 10.119600046984088
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The design of deep learning methods for low light video enhancement remains a
challenging problem owing to the difficulty in capturing low light and ground
truth video pairs. This is particularly hard in the context of dynamic scenes
or moving cameras where a long exposure ground truth cannot be captured. We
approach this problem by training a model on static videos such that the model
can generalize to dynamic videos. Existing methods adopting this approach
operate frame by frame and do not exploit the relationships among neighbouring
frames. We overcome this limitation through a selfcross dilated attention
module that can effectively learn to use information from neighbouring frames
even when dynamics between the frames are different during training and test
times. We validate our approach through experiments on multiple datasets and
show that our method outperforms other state-of-the-art video enhancement
algorithms when trained only on static videos.
Related papers
- ZeroSmooth: Training-free Diffuser Adaptation for High Frame Rate Video Generation [81.90265212988844]
We propose a training-free video method for generative video models in a plug-and-play manner.
We transform a video model into a self-cascaded video diffusion model with the designed hidden state correction modules.
Our training-free method is even comparable to trained models supported by huge compute resources and large-scale datasets.
arXiv Detail & Related papers (2024-06-03T00:31:13Z) - CTNeRF: Cross-Time Transformer for Dynamic Neural Radiance Field from Monocular Video [25.551944406980297]
We propose a novel approach to generate high-quality novel views from monocular videos of complex and dynamic scenes.
We introduce a module that operates in both the time and frequency domains to aggregate the features of object motion.
Our experiments demonstrate significant improvements over state-of-the-art methods on dynamic scene datasets.
arXiv Detail & Related papers (2024-01-10T00:40:05Z) - DynIBaR: Neural Dynamic Image-Based Rendering [79.44655794967741]
We address the problem of synthesizing novel views from a monocular video depicting a complex dynamic scene.
We adopt a volumetric image-based rendering framework that synthesizes new viewpoints by aggregating features from nearby views.
We demonstrate significant improvements over state-of-the-art methods on dynamic scene datasets.
arXiv Detail & Related papers (2022-11-20T20:57:02Z) - Video Frame Interpolation without Temporal Priors [91.04877640089053]
Video frame aims to synthesize non-exist intermediate frames in a video sequence.
The temporal priors of videos, i.e. frames per second (FPS) and frame exposure time, may vary from different camera sensors.
We devise a novel optical flow refinement strategy for better synthesizing results.
arXiv Detail & Related papers (2021-12-02T12:13:56Z) - Memory-Augmented Non-Local Attention for Video Super-Resolution [61.55700315062226]
We propose a novel video super-resolution method that aims at generating high-fidelity high-resolution (HR) videos from low-resolution (LR) ones.
Previous methods predominantly leverage temporal neighbor frames to assist the super-resolution of the current frame.
In contrast, we devise a cross-frame non-local attention mechanism that allows video super-resolution without frame alignment.
arXiv Detail & Related papers (2021-08-25T05:12:14Z) - Semi-Supervised Action Recognition with Temporal Contrastive Learning [50.08957096801457]
We learn a two-pathway temporal contrastive model using unlabeled videos at two different speeds.
We considerably outperform video extensions of sophisticated state-of-the-art semi-supervised image recognition methods.
arXiv Detail & Related papers (2021-02-04T17:28:35Z) - Video Deblurring by Fitting to Test Data [39.41334067434719]
Motion blur in videos captured by autonomous vehicles and robots can degrade their perception capability.
We present a novel approach to video deblurring by fitting a deep network to the test video.
Our approach selects sharp frames from a video and then trains a convolutional neural network on these sharp frames.
arXiv Detail & Related papers (2020-12-09T18:49:24Z) - Deep Motion Blind Video Stabilization [4.544151613454639]
This work aims to declutter this over-complicated formulation of video stabilization with the help of a novel dataset.
We successfully learn motion blind full-frame video stabilization through employing strictly conventional generative techniques.
Our method achieves $sim3times$ speed-up over the currently available fastest video stabilization methods.
arXiv Detail & Related papers (2020-11-19T07:26:06Z) - Non-Adversarial Video Synthesis with Learned Priors [53.26777815740381]
We focus on the problem of generating videos from latent noise vectors, without any reference input frames.
We develop a novel approach that jointly optimize the input latent space, the weights of a recurrent neural network and a generator through non-adversarial learning.
Our approach generates superior quality videos compared to the existing state-of-the-art methods.
arXiv Detail & Related papers (2020-03-21T02:57:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.