Multi-modal Datasets for Super-resolution
- URL: http://arxiv.org/abs/2004.05804v1
- Date: Mon, 13 Apr 2020 07:39:52 GMT
- Title: Multi-modal Datasets for Super-resolution
- Authors: Haoran Li, Weihong Quan, Meijun Yan, Jin zhang, Xiaoli Gong and Jin
Zhou
- Abstract summary: We propose real-world black-and-white old photo datasets for super-resolution (OID-RW)
The dataset contains 82 groups of images, including 22 groups of character type and 60 groups of landscape and architecture.
We also propose a multi-modal degradation dataset (MDD400) to solve the super-resolution reconstruction in real-life image degradation scenarios.
- Score: 12.079245552387361
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Nowdays, most datasets used to train and evaluate super-resolution models are
single-modal simulation datasets. However, due to the variety of image
degradation types in the real world, models trained on single-modal simulation
datasets do not always have good robustness and generalization ability in
different degradation scenarios. Previous work tended to focus only on
true-color images. In contrast, we first proposed real-world black-and-white
old photo datasets for super-resolution (OID-RW), which is constructed using
two methods of manually filling pixels and shooting with different cameras. The
dataset contains 82 groups of images, including 22 groups of character type and
60 groups of landscape and architecture. At the same time, we also propose a
multi-modal degradation dataset (MDD400) to solve the super-resolution
reconstruction in real-life image degradation scenarios. We managed to simulate
the process of generating degraded images by the following four methods:
interpolation algorithm, CNN network, GAN network and capturing videos with
different bit rates. Our experiments demonstrate that not only the models
trained on our dataset have better generalization capability and robustness,
but also the trained images can maintain better edge contours and texture
features.
Related papers
- Exposure Bracketing is All You Need for Unifying Image Restoration and Enhancement Tasks [50.822601495422916]
We propose to utilize exposure bracketing photography to unify image restoration and enhancement tasks.
Due to the difficulty in collecting real-world pairs, we suggest a solution that first pre-trains the model with synthetic paired data.
In particular, a temporally modulated recurrent network (TMRNet) and self-supervised adaptation method are proposed.
arXiv Detail & Related papers (2024-01-01T14:14:35Z) - Towards Real-World Burst Image Super-Resolution: Benchmark and Method [93.73429028287038]
In this paper, we establish a large-scale real-world burst super-resolution dataset, i.e., RealBSR, to explore the faithful reconstruction of image details from multiple frames.
We also introduce a Federated Burst Affinity network (FBAnet) to investigate non-trivial pixel-wise displacement among images under real-world image degradation.
arXiv Detail & Related papers (2023-09-09T14:11:37Z) - ParaColorizer: Realistic Image Colorization using Parallel Generative
Networks [1.7778609937758327]
Grayscale image colorization is a fascinating application of AI for information restoration.
We present a parallel GAN-based colorization framework.
We show the shortcomings of the non-perceptual evaluation metrics commonly used to assess multi-modal problems.
arXiv Detail & Related papers (2022-08-17T13:49:44Z) - Learning Enriched Features for Fast Image Restoration and Enhancement [166.17296369600774]
This paper presents a holistic goal of maintaining spatially-precise high-resolution representations through the entire network.
We learn an enriched set of features that combines contextual information from multiple scales, while simultaneously preserving the high-resolution spatial details.
Our approach achieves state-of-the-art results for a variety of image processing tasks, including defocus deblurring, image denoising, super-resolution, and image enhancement.
arXiv Detail & Related papers (2022-04-19T17:59:45Z) - Any-resolution Training for High-resolution Image Synthesis [55.19874755679901]
Generative models operate at fixed resolution, even though natural images come in a variety of sizes.
We argue that every pixel matters and create datasets with variable-size images, collected at their native resolutions.
We introduce continuous-scale training, a process that samples patches at random scales to train a new generator with variable output resolutions.
arXiv Detail & Related papers (2022-04-14T17:59:31Z) - Plotting time: On the usage of CNNs for time series classification [1.0390583509657398]
We present a novel approach for time series classification where we represent time series data as plot images and feed them to a simple CNN.
Our approach is very promising, achieving the best results on both real-world datasets and matching / beating the best state-of-the-art methods in six UCR datasets.
arXiv Detail & Related papers (2021-02-08T13:23:01Z) - Exploiting Raw Images for Real-Scene Super-Resolution [105.18021110372133]
We study the problem of real-scene single image super-resolution to bridge the gap between synthetic data and real captured images.
We propose a method to generate more realistic training data by mimicking the imaging process of digital cameras.
We also develop a two-branch convolutional neural network to exploit the radiance information originally-recorded in raw images.
arXiv Detail & Related papers (2021-02-02T16:10:15Z) - Bayesian Image Reconstruction using Deep Generative Models [7.012708932320081]
In this work, we leverage state-of-the-art (SOTA) generative models for building powerful image priors.
Our method, called Bayesian Reconstruction through Generative Models (BRGM), uses a single pre-trained generator model to solve different image restoration tasks.
arXiv Detail & Related papers (2020-12-08T17:11:26Z) - Learning Enriched Features for Real Image Restoration and Enhancement [166.17296369600774]
convolutional neural networks (CNNs) have achieved dramatic improvements over conventional approaches for image restoration task.
We present a novel architecture with the collective goals of maintaining spatially-precise high-resolution representations through the entire network.
Our approach learns an enriched set of features that combines contextual information from multiple scales, while simultaneously preserving the high-resolution spatial details.
arXiv Detail & Related papers (2020-03-15T11:04:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.