Gated Recurrent Unit for Video Denoising
- URL: http://arxiv.org/abs/2210.09135v1
- Date: Mon, 17 Oct 2022 14:34:54 GMT
- Title: Gated Recurrent Unit for Video Denoising
- Authors: Kai Guo, Seungwon Choi and Jongseong Choi
- Abstract summary: We propose a new video denoising model based on gated recurrent unit (GRU) mechanisms for video denoising.
The experimental results show that the GRU-VD network can achieve better quality than state of the arts objectively and subjectively.
- Score: 5.515903319513226
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Current video denoising methods perform temporal fusion by designing
convolutional neural networks (CNN) or combine spatial denoising with temporal
fusion into basic recurrent neural networks (RNNs). However, there have not yet
been works which adapt gated recurrent unit (GRU) mechanisms for video
denoising. In this letter, we propose a new video denoising model based on GRU,
namely GRU-VD. First, the reset gate is employed to mark the content related to
the current frame in the previous frame output. Then the hidden activation
works as an initial spatial-temporal denoising with the help from the marked
relevant content. Finally, the update gate recursively fuses the initial
denoised result with previous frame output to further increase accuracy. To
handle various light conditions adaptively, the noise standard deviation of the
current frame is also fed to these three modules. A weighted loss is adopted to
regulate initial denoising and final fusion at the same time. The experimental
results show that the GRU-VD network not only can achieve better quality than
state of the arts objectively and subjectively, but also can obtain satisfied
subjective quality on real video.
Related papers
- A Practical Gated Recurrent Transformer Network Incorporating Multiple Fusions for Video Denoising [1.5044085747326295]
State-of-the-art (SOTA) video denoising methods employ multi-frame simultaneous denoising mechanisms.
We propose a multi-fusion gated recurrent Transformer network (GRTN) that achieves SOTA denoising performance with only a single-frame delay.
arXiv Detail & Related papers (2024-09-10T15:55:53Z) - Boosting Neural Representations for Videos with a Conditional Decoder [28.073607937396552]
Implicit neural representations (INRs) have emerged as a promising approach for video storage and processing.
This paper introduces a universal boosting framework for current implicit video representation approaches.
arXiv Detail & Related papers (2024-02-28T08:32:19Z) - Video Dynamics Prior: An Internal Learning Approach for Robust Video
Enhancements [83.5820690348833]
We present a framework for low-level vision tasks that does not require any external training data corpus.
Our approach learns neural modules by optimizing over a corrupted sequence, leveraging the weights of the coherence-temporal test and statistics internal statistics.
arXiv Detail & Related papers (2023-12-13T01:57:11Z) - No-frills Temporal Video Grounding: Multi-Scale Neighboring Attention
and Zoom-in Boundary Detection [52.03562682785128]
Temporal video grounding aims to retrieve the time interval of a language query from an untrimmed video.
A significant challenge in TVG is the low "Semantic Noise Ratio (SNR)", which results in worse performance with lower SNR.
We propose a no-frills TVG model that consists of two core modules, namely multi-scale neighboring attention and zoom-in boundary detection.
arXiv Detail & Related papers (2023-07-20T04:12:10Z) - Real-time Controllable Denoising for Image and Video [44.68523669975698]
Controllable image denoising aims to generate clean samples with human priors and balance sharpness and smoothness.
We introduce Real-time Controllable Denoising (RCD), the first deep image and video denoising pipeline.
RCD provides a fully controllable user interface to edit arbitrary denoising levels in real-time with only one-time network inference.
arXiv Detail & Related papers (2023-03-29T03:10:28Z) - Making Reconstruction-based Method Great Again for Video Anomaly
Detection [64.19326819088563]
Anomaly detection in videos is a significant yet challenging problem.
Existing reconstruction-based methods rely on old-fashioned convolutional autoencoders.
We propose a new autoencoder model for enhanced consecutive frame reconstruction.
arXiv Detail & Related papers (2023-01-28T01:57:57Z) - Multi-stage image denoising with the wavelet transform [125.2251438120701]
Deep convolutional neural networks (CNNs) are used for image denoising via automatically mining accurate structure information.
We propose a multi-stage image denoising CNN with the wavelet transform (MWDCNN) via three stages, i.e., a dynamic convolutional block (DCB), two cascaded wavelet transform and enhancement blocks (WEBs) and residual block (RB)
arXiv Detail & Related papers (2022-09-26T03:28:23Z) - Learning Task-Oriented Flows to Mutually Guide Feature Alignment in
Synthesized and Real Video Denoising [137.5080784570804]
Video denoising aims at removing noise from videos to recover clean ones.
Some existing works show that optical flow can help the denoising by exploiting the additional spatial-temporal clues from nearby frames.
We propose a new multi-scale refined optical flow-guided video denoising method, which is more robust to different noise levels.
arXiv Detail & Related papers (2022-08-25T00:09:18Z) - STIP: A SpatioTemporal Information-Preserving and Perception-Augmented
Model for High-Resolution Video Prediction [78.129039340528]
We propose a Stemporal Information-Preserving and Perception-Augmented Model (STIP) to solve the above two problems.
The proposed model aims to preserve thetemporal information for videos during the feature extraction and the state transitions.
Experimental results show that the proposed STIP can predict videos with more satisfactory visual quality compared with a variety of state-of-the-art methods.
arXiv Detail & Related papers (2022-06-09T09:49:04Z) - Enhancement of a CNN-Based Denoiser Based on Spatial and Spectral
Analysis [23.11994688706024]
We propose a discrete wavelet denoising CNN (WDnCNN) which restores images corrupted by various noise with a single model.
To address this issue, we present a band normalization module (BNM) to normalize the coefficients from different parts of the frequency spectrum.
We evaluate the proposed WDnCNN, and compare it with other state-of-the-art denoisers.
arXiv Detail & Related papers (2020-06-28T05:25:50Z) - Restore from Restored: Video Restoration with Pseudo Clean Video [28.057705167363327]
We propose a self-supervised video denoising method called "restore-from-restored"
This method fine-tunes a pre-trained network by using a pseudo clean video during the test phase.
We analyze the restoration performance of the fine-tuned video denoising networks with the proposed self-supervision-based learning algorithm.
arXiv Detail & Related papers (2020-03-09T17:37:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.