HQ-50K: A Large-scale, High-quality Dataset for Image Restoration
- URL: http://arxiv.org/abs/2306.05390v1
- Date: Thu, 8 Jun 2023 17:44:21 GMT
- Title: HQ-50K: A Large-scale, High-quality Dataset for Image Restoration
- Authors: Qinhong Yang and Dongdong Chen and Zhentao Tan and Qiankun Liu and Qi
Chu and Jianmin Bao and Lu Yuan and Gang Hua and Nenghai Yu
- Abstract summary: HQ-50K contains 50,000 high-quality images with rich texture details and semantic diversity.
We analyze existing image restoration datasets from five different perspectives.
HQ-50K considers all of these five aspects during the data curation process and meets all requirements.
- Score: 105.22191357934398
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper introduces a new large-scale image restoration dataset, called
HQ-50K, which contains 50,000 high-quality images with rich texture details and
semantic diversity. We analyze existing image restoration datasets from five
different perspectives, including data scale, resolution, compression rates,
texture details, and semantic coverage. However, we find that all of these
datasets are deficient in some aspects. In contrast, HQ-50K considers all of
these five aspects during the data curation process and meets all requirements.
We also present a new Degradation-Aware Mixture of Expert (DAMoE) model, which
enables a single model to handle multiple corruption types and unknown levels.
Our extensive experiments demonstrate that HQ-50K consistently improves the
performance on various image restoration tasks, such as super-resolution,
denoising, dejpeg, and deraining. Furthermore, our proposed DAMoE, trained on
our \dataset, outperforms existing state-of-the-art unified models designed for
multiple restoration tasks and levels. The dataset and code are available at
\url{https://github.com/littleYaang/HQ-50K}.
Related papers
- DocIQ: A Benchmark Dataset and Feature Fusion Network for Document Image Quality Assessment [78.21680156380705]
We introduce a subjective DIQA dataset DIQA-5000.<n>The DIQA-5000 dataset comprises 5,000 document images, generated by applying multiple document enhancement techniques to 500 real-world images with diverse distortions.<n>Each enhanced image was rated by 15 subjects across three rating dimensions: overall quality, sharpness, and color fidelity.<n>We propose a specialized no-reference DIQA model that exploits document layout features to maintain quality perception at reduced resolutions to lower computational cost.
arXiv Detail & Related papers (2025-09-21T10:01:43Z) - No Pixel Left Behind: A Detail-Preserving Architecture for Robust High-Resolution AI-Generated Image Detection [15.139983859649922]
High-Resolution Detail-Aggregation Network (HiDA-Net) is a novel framework that ensures no pixel is left behind.<n>HiDA-Net achieves state-of-the-art, increasing accuracy by over 13% on the challenging Chameleon dataset and 10% on our HiRes-50K.
arXiv Detail & Related papers (2025-08-24T13:03:16Z) - OpenRR-5k: A Large-Scale Benchmark for Reflection Removal in the Wild [11.90005156583499]
We present a novel benchmark for Single Image Reflection Removal (SIRR)<n>We have developed a large-scale dataset containing 5,300 high-quality, pixel-aligned image pairs.
arXiv Detail & Related papers (2025-06-05T18:03:39Z) - Rethinking Image Super-Resolution from Training Data Perspectives [54.28824316574355]
We investigate the understudied effect of the training data used for image super-resolution (SR)
With this, we propose an automated image evaluation pipeline.
We find that datasets with (i) low compression artifacts, (ii) high within-image diversity as judged by the number of different objects, and (iii) a large number of images from ImageNet or PASS all positively affect SR performance.
arXiv Detail & Related papers (2024-09-01T16:25:04Z) - Review Learning: Advancing All-in-One Ultra-High-Definition Image Restoration Training Method [7.487270862599671]
We propose a new training paradigm for general image restoration models, which we name bfReview Learning.
This approach begins with sequential training of an image restoration model on several degraded datasets, combined with a review mechanism.
We design a lightweight all-purpose image restoration network that can efficiently reason about degraded images with 4K resolution on a single consumer-grade GPU.
arXiv Detail & Related papers (2024-08-13T08:08:45Z) - UHD-IQA Benchmark Database: Pushing the Boundaries of Blind Photo Quality Assessment [4.563959812257119]
We introduce a novel Image Quality Assessment dataset comprising 6073 UHD-1 (4K) images, annotated at a fixed width of 3840 pixels.
Ours focuses on highly aesthetic photos of high technical quality, filling a gap in the literature.
The dataset is annotated with perceptual quality ratings obtained through a crowdsourcing study.
arXiv Detail & Related papers (2024-06-25T11:30:31Z) - InstructIR: High-Quality Image Restoration Following Human Instructions [61.1546287323136]
We present the first approach that uses human-written instructions to guide the image restoration model.
Our method, InstructIR, achieves state-of-the-art results on several restoration tasks.
arXiv Detail & Related papers (2024-01-29T18:53:33Z) - Self-Supervised Masked Digital Elevation Models Encoding for
Low-Resource Downstream Tasks [0.6374763930914523]
GeoAI is uniquely poised to take advantage of the self-supervised methodology due to the decades of data collected.
The proposed architecture is the Masked Autoencoder pre-trained on ImageNet.
arXiv Detail & Related papers (2023-09-06T21:20:10Z) - A Comprehensive Study of Modern Architectures and Regularization
Approaches on CheXpert5000 [3.7384509727711923]
We present a study on medical image classification with limited annotations (5k)
We find that models pretrained on ImageNet21k achieve a higher AUC and larger models require less training steps.
Vision Transformer achieve comparable or on par results to Big Transfer Models.
arXiv Detail & Related papers (2023-02-13T20:51:24Z) - Highly Accurate Dichotomous Image Segmentation [139.79513044546]
A new task called dichotomous image segmentation (DIS) aims to segment highly accurate objects from natural images.
We collect the first large-scale dataset, DIS5K, which contains 5,470 high-resolution (e.g., 2K, 4K or larger) images.
We also introduce a simple intermediate supervision baseline (IS-Net) using both feature-level and mask-level guidance for DIS model training.
arXiv Detail & Related papers (2022-03-06T20:09:19Z) - Salient Objects in Clutter [130.63976772770368]
This paper identifies and addresses a serious design bias of existing salient object detection (SOD) datasets.
This design bias has led to a saturation in performance for state-of-the-art SOD models when evaluated on existing datasets.
We propose a new high-quality dataset and update the previous saliency benchmark.
arXiv Detail & Related papers (2021-05-07T03:49:26Z) - Multi-Stage Progressive Image Restoration [167.6852235432918]
We propose a novel synergistic design that can optimally balance these competing goals.
Our main proposal is a multi-stage architecture, that progressively learns restoration functions for the degraded inputs.
The resulting tightly interlinked multi-stage architecture, named as MPRNet, delivers strong performance gains on ten datasets.
arXiv Detail & Related papers (2021-02-04T18:57:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.