Benchmarking Ultra-High-Definition Image Reflection Removal
- URL: http://arxiv.org/abs/2308.00265v1
- Date: Tue, 1 Aug 2023 03:56:50 GMT
- Title: Benchmarking Ultra-High-Definition Image Reflection Removal
- Authors: Zhenyuan Zhang, Zhenbo Song, Kaihao Zhang, Wenhan Luo, Zhaoxin Fan,
Jianfeng Lu
- Abstract summary: We propose a transformer-based architecture named RRFormer for reflection removal.
We show that RRFormer achieves state-of-the-art performance on both the non-UHD dataset and our proposed UHDRR datasets.
- Score: 43.34342640578687
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep learning based methods have achieved significant success in the task of
single image reflection removal (SIRR). However, the majority of these methods
are focused on High-Definition/Standard-Definition (HD/SD) images, while
ignoring higher resolution images such as Ultra-High-Definition (UHD) images.
With the increasing prevalence of UHD images captured by modern devices, in
this paper, we aim to address the problem of UHD SIRR. Specifically, we first
synthesize two large-scale UHD datasets, UHDRR4K and UHDRR8K. The UHDRR4K
dataset consists of $2,999$ and $168$ quadruplets of images for training and
testing respectively, and the UHDRR8K dataset contains $1,014$ and $105$
quadruplets. To the best of our knowledge, these two datasets are the first
largest-scale UHD datasets for SIRR. Then, we conduct a comprehensive
evaluation of six state-of-the-art SIRR methods using the proposed datasets.
Based on the results, we provide detailed discussions regarding the strengths
and limitations of these methods when applied to UHD images. Finally, we
present a transformer-based architecture named RRFormer for reflection removal.
RRFormer comprises three modules, namely the Prepossessing Embedding Module,
Self-attention Feature Extraction Module, and Multi-scale Spatial Feature
Extraction Module. These modules extract hypercolumn features, global and
partial attention features, and multi-scale spatial features, respectively. To
ensure effective training, we utilize three terms in our loss function: pixel
loss, feature loss, and adversarial loss. We demonstrate through experimental
results that RRFormer achieves state-of-the-art performance on both the non-UHD
dataset and our proposed UHDRR datasets. The code and datasets are publicly
available at
https://github.com/Liar-zzy/Benchmarking-Ultra-High-Definition-Single-Image-Reflection-Removal.
Related papers
- Ultra-High-Definition Restoration: New Benchmarks and A Dual Interaction Prior-Driven Solution [35.58381343948639]
We construct UHD snow and rain benchmarks, named UHD-Snow and UHD-Rain.
Each benchmark contains 3200 degraded/clear image pairs of 4K resolution.
We propose an effective UHD image restoration solution by considering gradient and normal priors in model design.
arXiv Detail & Related papers (2024-06-19T14:58:49Z) - Towards Ultra-High-Definition Image Deraining: A Benchmark and An Efficient Method [42.331058889312466]
This paper contributes the first large-scale UHD image deraining dataset, 4K-Rain13k, that contains 13,000 image pairs at 4K resolution.
We develop an effective and efficient vision-based architecture (UDR-Mixer) to better solve this task.
arXiv Detail & Related papers (2024-05-27T11:45:08Z) - Generating Content for HDR Deghosting from Frequency View [56.103761824603644]
Recent Diffusion Models (DMs) have been introduced in HDR imaging field.
DMs require extensive iterations with large models to estimate entire images.
We propose the Low-Frequency aware Diffusion (LF-Diff) model for ghost-free HDR imaging.
arXiv Detail & Related papers (2024-04-01T01:32:11Z) - RawHDR: High Dynamic Range Image Reconstruction from a Single Raw Image [36.17182977927645]
High dynamic range (RGB) images capture much more intensity levels than standard ones.
Current methods predominantly generate HDR images from 8-bit low dynamic range (LDR) s images that have been degraded by the camera processing pipeline.
Unlike existing methods, the core idea of this work is to incorporate more informative Raw sensor data to generate HDR images.
arXiv Detail & Related papers (2023-09-05T07:58:21Z) - Recurrent Multi-scale Transformer for High-Resolution Salient Object
Detection [68.65338791283298]
Salient Object Detection (SOD) aims to identify and segment the most conspicuous objects in an image or video.
Traditional SOD methods are largely limited to low-resolution images, making them difficult to adapt to the development of High-Resolution SOD.
In this work, we first propose a new HRS10K dataset, which contains 10,500 high-quality annotated images at 2K-8K resolution.
arXiv Detail & Related papers (2023-08-07T17:49:04Z) - Embedding Fourier for Ultra-High-Definition Low-Light Image Enhancement [78.67036949708795]
Ultra-High-Definition (UHD) photo has gradually become the standard configuration in advanced imaging devices.
We propose a new solution, UHDFour, that embeds Fourier transform into a cascaded network.
We also contribute the first real UHD LLIE dataset, textbfUHD-LL, that contains 2,150 low-noise/normal-clear 4K image pairs.
arXiv Detail & Related papers (2023-02-23T07:43:41Z) - A Two-stage Deep Network for High Dynamic Range Image Reconstruction [0.883717274344425]
This study tackles the challenges of single-shot LDR to HDR mapping by proposing a novel two-stage deep network.
Notably, our proposed method aims to reconstruct an HDR image without knowing hardware information, including camera response function (CRF) and exposure settings.
arXiv Detail & Related papers (2021-04-19T15:19:17Z) - WHU-Hi: UAV-borne hyperspectral with high spatial resolution (H2)
benchmark datasets for hyperspectral image classification [5.281167336437183]
A new benchmark dataset named the Wuhan UAV-borne hyperspectral image (WHU-Hi) dataset was built for hyperspectral image classification.
The WHU-Hi dataset has a high spectral resolution (nm level) and a very high spatial resolution (cm level)
Some start-of-art hyperspectral image classification methods benchmarked the WHU-Hi dataset, and the experimental results show that WHU-Hi is a challenging dataset.
arXiv Detail & Related papers (2020-12-27T11:28:37Z) - HDR-GAN: HDR Image Reconstruction from Multi-Exposed LDR Images with
Large Motions [62.44802076971331]
We propose a novel GAN-based model, HDR-GAN, for synthesizing HDR images from multi-exposed LDR images.
By incorporating adversarial learning, our method is able to produce faithful information in the regions with missing content.
arXiv Detail & Related papers (2020-07-03T11:42:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.