Self-supervision versus synthetic datasets: which is the lesser evil in
the context of video denoising?
- URL: http://arxiv.org/abs/2204.11493v1
- Date: Mon, 25 Apr 2022 08:17:36 GMT
- Title: Self-supervision versus synthetic datasets: which is the lesser evil in
the context of video denoising?
- Authors: Val\'ery Dewil, Aranud Barral, Gabriele Facciolo, Pablo Arias
- Abstract summary: Supervised training has led to state-of-the-art results in image and video denoising.
It requires large datasets of noisy-clean pairs that are difficult to obtain.
Some self-supervised frameworks have been proposed for training such denoising networks directly on the noisy data.
- Score: 11.0189148044343
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Supervised training has led to state-of-the-art results in image and video
denoising. However, its application to real data is limited since it requires
large datasets of noisy-clean pairs that are difficult to obtain. For this
reason, networks are often trained on realistic synthetic data. More recently,
some self-supervised frameworks have been proposed for training such denoising
networks directly on the noisy data without requiring ground truth. On
synthetic denoising problems supervised training outperforms self-supervised
approaches, however in recent years the gap has become narrower, especially for
video. In this paper, we propose a study aiming to determine which is the best
approach to train denoising networks for real raw videos: supervision on
synthetic realistic data or self-supervision on real data. A complete study
with quantitative results in case of natural videos with real motion is
impossible since no dataset with clean-noisy pairs exists. We address this
issue by considering three independent experiments in which we compare the two
frameworks. We found that self-supervision on the real data outperforms
supervision on synthetic data, and that in normal illumination conditions the
drop in performance is due to the synthetic ground truth generation, not the
noise model.
Related papers
- Video DataFlywheel: Resolving the Impossible Data Trinity in Video-Language Understanding [61.89781979702939]
This study quantitatively reveals an "impossible trinity" among data quantity, diversity, and quality in pre-training datasets.
Recent efforts seek to refine large-scale, diverse ASR datasets compromised by low quality through synthetic annotations.
We introduce the Video DataFlywheel framework, which iteratively refines video annotations with improved noise control methods.
arXiv Detail & Related papers (2024-09-29T03:33:35Z) - Learning with Noisy Foundation Models [95.50968225050012]
This paper is the first work to comprehensively understand and analyze the nature of noise in pre-training datasets.
We propose a tuning method (NMTune) to affine the feature space to mitigate the malignant effect of noise and improve generalization.
arXiv Detail & Related papers (2024-03-11T16:22:41Z) - Learning Defect Prediction from Unrealistic Data [57.53586547895278]
Pretrained models of code have become popular choices for code understanding and generation tasks.
Such models tend to be large and require commensurate volumes of training data.
It has become popular to train models with far larger but less realistic datasets, such as functions with artificially injected bugs.
Models trained on such data tend to only perform well on similar data, while underperforming on real world programs.
arXiv Detail & Related papers (2023-11-02T01:51:43Z) - Double Descent and Overfitting under Noisy Inputs and Distribution Shift for Linear Denoisers [3.481985817302898]
A concern about studying supervised denoising is that one might not always have noiseless training data from the test distribution.
Motivated by this, we study supervised denoising and noisy-input regression under distribution shift.
arXiv Detail & Related papers (2023-05-26T22:41:40Z) - Realistic Noise Synthesis with Diffusion Models [68.48859665320828]
Deep image denoising models often rely on large amount of training data for the high quality performance.
We propose a novel method that synthesizes realistic noise using diffusion models, namely Realistic Noise Synthesize Diffusor (RNSD)
RNSD can incorporate guided multiscale content, such as more realistic noise with spatial correlations can be generated at multiple frequencies.
arXiv Detail & Related papers (2023-05-23T12:56:01Z) - Confidence-based Reliable Learning under Dual Noises [46.45663546457154]
Deep neural networks (DNNs) have achieved remarkable success in a variety of computer vision tasks.
Yet, the data collected from the open world are unavoidably polluted by noise, which may significantly undermine the efficacy of the learned models.
Various attempts have been made to reliably train DNNs under data noise, but they separately account for either the noise existing in the labels or that existing in the images.
This work provides a first, unified framework for reliable learning under the joint (image, label)-noise.
arXiv Detail & Related papers (2023-02-10T07:50:34Z) - IDR: Self-Supervised Image Denoising via Iterative Data Refinement [66.5510583957863]
We present a practical unsupervised image denoising method to achieve state-of-the-art denoising performance.
Our method only requires single noisy images and a noise model, which is easily accessible in practical raw image denoising.
To evaluate raw image denoising performance in real-world applications, we build a high-quality raw image dataset SenseNoise-500 that contains 500 real-life scenes.
arXiv Detail & Related papers (2021-11-29T07:22:53Z) - Semi-synthesis: A fast way to produce effective datasets for stereo
matching [16.602343511350252]
Close-to-real-scene texture rendering is a key factor to boost up stereo matching performance.
We propose semi-synthetic, an effective and fast way to synthesize large amount of data with close-to-real-scene texture.
With further fine-tuning on the real dataset, we also achieve SOTA performance on Middlebury and competitive results on KITTI and ETH3D datasets.
arXiv Detail & Related papers (2021-01-26T14:34:49Z) - Joint Demosaicking and Denoising in the Wild: The Case of Training Under
Ground Truth Uncertainty [2.0088802641040604]
We propose and study Wild-JDD, a novel learning framework for joint demosaicking and denoising in the wild.
In contrast to previous works which generally assume the ground truth of training data is a perfect reflection of the reality, we consider here the more common imperfect case of ground truth uncertainty in the wild.
arXiv Detail & Related papers (2021-01-12T12:33:41Z) - CycleISP: Real Image Restoration via Improved Data Synthesis [166.17296369600774]
We present a framework that models camera imaging pipeline in forward and reverse directions.
By training a new image denoising network on realistic synthetic data, we achieve the state-of-the-art performance on real camera benchmark datasets.
arXiv Detail & Related papers (2020-03-17T15:20:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.