First image then video: A two-stage network for spatiotemporal video
denoising
- URL: http://arxiv.org/abs/2001.00346v2
- Date: Wed, 22 Jan 2020 03:36:15 GMT
- Title: First image then video: A two-stage network for spatiotemporal video
denoising
- Authors: Ce Wang, S. Kevin Zhou, Zhiwei Cheng
- Abstract summary: Video denoising is to remove noise from noise-corrupted data, thus recovering true motion signals.
Existing approaches for video denoising tend to suffer from blur artifacts, that is the boundary of a moving object tends to appear blurry.
This paper introduces a first-image-then-video two-stage denoising neural network, consisting of an image denoising module and a regular intratemporal video denoising module.
It yields state-of-the-art performances on the video denoising Vimeo90K dataset in terms of both denoising quality and computation.
- Score: 19.842488445174524
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Video denoising is to remove noise from noise-corrupted data, thus recovering
true signals via spatiotemporal processing. Existing approaches for
spatiotemporal video denoising tend to suffer from motion blur artifacts, that
is, the boundary of a moving object tends to appear blurry especially when the
object undergoes a fast motion, causing optical flow calculation to break down.
In this paper, we address this challenge by designing a first-image-then-video
two-stage denoising neural network, consisting of an image denoising module for
spatially reducing intra-frame noise followed by a regular spatiotemporal video
denoising module. The intuition is simple yet powerful and effective: the first
stage of image denoising effectively reduces the noise level and, therefore,
allows the second stage of spatiotemporal denoising for better modeling and
learning everywhere, including along the moving object boundaries. This
two-stage network, when trained in an end-to-end fashion, yields the
state-of-the-art performances on the video denoising benchmark Vimeo90K dataset
in terms of both denoising quality and computation. It also enables an
unsupervised approach that achieves comparable performance to existing
supervised approaches.
Related papers
- Temporal As a Plugin: Unsupervised Video Denoising with Pre-Trained Image Denoisers [30.965705043127144]
In this paper, we propose a novel unsupervised video denoising framework, named Temporal As aTAP' (TAP)
By incorporating temporal modules, our method can harness temporal information across noisy frames, complementing its power of spatial denoising.
Compared to other unsupervised video denoising methods, our framework demonstrates superior performance on both sRGB and raw video denoising datasets.
arXiv Detail & Related papers (2024-09-17T15:05:33Z) - TRIP: Temporal Residual Learning with Image Noise Prior for Image-to-Video Diffusion Models [94.24861019513462]
TRIP is a new recipe of image-to-video diffusion paradigm.
It pivots on image noise prior derived from static image to jointly trigger inter-frame relational reasoning.
Extensive experiments on WebVid-10M, DTDB and MSR-VTT datasets demonstrate TRIP's effectiveness.
arXiv Detail & Related papers (2024-03-25T17:59:40Z) - Real-time Controllable Denoising for Image and Video [44.68523669975698]
Controllable image denoising aims to generate clean samples with human priors and balance sharpness and smoothness.
We introduce Real-time Controllable Denoising (RCD), the first deep image and video denoising pipeline.
RCD provides a fully controllable user interface to edit arbitrary denoising levels in real-time with only one-time network inference.
arXiv Detail & Related papers (2023-03-29T03:10:28Z) - Low Latency Video Denoising for Online Conferencing Using CNN
Architectures [4.7805617044617446]
We propose a pipeline for real-time video denoising with low runtime cost and high perceptual quality.
A custom noise detector analyzer provides real-time feedback to adapt the weights and improve the models' output.
arXiv Detail & Related papers (2023-02-17T00:55:54Z) - Real-time Streaming Video Denoising with Bidirectional Buffers [48.57108807146537]
Real-time denoising algorithms are typically adopted on the user device to remove the noise involved during the shooting and transmission of video streams.
Recent multi-output inference works propagate the bidirectional temporal feature with a parallel or recurrent framework.
We propose a Bidirectional Streaming Video Denoising framework, to achieve high-fidelity real-time denoising for streaming videos with both past and future temporal receptive fields.
arXiv Detail & Related papers (2022-07-14T14:01:03Z) - IDR: Self-Supervised Image Denoising via Iterative Data Refinement [66.5510583957863]
We present a practical unsupervised image denoising method to achieve state-of-the-art denoising performance.
Our method only requires single noisy images and a noise model, which is easily accessible in practical raw image denoising.
To evaluate raw image denoising performance in real-world applications, we build a high-quality raw image dataset SenseNoise-500 that contains 500 real-life scenes.
arXiv Detail & Related papers (2021-11-29T07:22:53Z) - Learning Spatial and Spatio-Temporal Pixel Aggregations for Image and
Video Denoising [104.59305271099967]
We present a pixel aggregation network and learn the pixel sampling and averaging strategies for image denoising.
We develop a pixel aggregation network for video denoising to sample pixels across the spatial-temporal space.
Our method is able to solve the misalignment issues caused by large motion in dynamic scenes.
arXiv Detail & Related papers (2021-01-26T13:00:46Z) - Neighbor2Neighbor: Self-Supervised Denoising from Single Noisy Images [98.82804259905478]
We present Neighbor2Neighbor to train an effective image denoising model with only noisy images.
In detail, input and target used to train a network are images sub-sampled from the same noisy image.
A denoising network is trained on sub-sampled training pairs generated in the first stage, with a proposed regularizer as additional loss for better performance.
arXiv Detail & Related papers (2021-01-08T02:03:25Z) - Learning Model-Blind Temporal Denoisers without Ground Truths [46.778450578529814]
Denoisers trained with synthetic data often fail to cope with the diversity of unknown noises.
Previous image-based method leads to noise overfitting if directly applied to video denoisers.
We propose a general framework for video denoising networks that successfully addresses these challenges.
arXiv Detail & Related papers (2020-07-07T07:19:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.