Towards Efficient and Scale-Robust Ultra-High-Definition Image
Demoireing
- URL: http://arxiv.org/abs/2207.09935v1
- Date: Wed, 20 Jul 2022 14:20:52 GMT
- Title: Towards Efficient and Scale-Robust Ultra-High-Definition Image
Demoireing
- Authors: Xin Yu, Peng Dai, Wenbo Li, Lan Ma, Jiajun Shen, Jia Li, Xiaojuan Qi
- Abstract summary: We present an efficient baseline model ESDNet for tackling 4K moire images, wherein we build a semantic-aligned scale-aware module to address the scale variation of moire patterns.
Our approach outperforms state-of-the-art methods by a large margin while being much more lightweight.
- Score: 71.62289021118983
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: With the rapid development of mobile devices, modern widely-used mobile
phones typically allow users to capture 4K resolution (i.e.,
ultra-high-definition) images. However, for image demoireing, a challenging
task in low-level vision, existing works are generally carried out on
low-resolution or synthetic images. Hence, the effectiveness of these methods
on 4K resolution images is still unknown. In this paper, we explore moire
pattern removal for ultra-high-definition images. To this end, we propose the
first ultra-high-definition demoireing dataset (UHDM), which contains 5,000
real-world 4K resolution image pairs, and conduct a benchmark study on current
state-of-the-art methods. Further, we present an efficient baseline model
ESDNet for tackling 4K moire images, wherein we build a semantic-aligned
scale-aware module to address the scale variation of moire patterns. Extensive
experiments manifest the effectiveness of our approach, which outperforms
state-of-the-art methods by a large margin while being much more lightweight.
Code and dataset are available at https://xinyu-andy.github.io/uhdm-page.
Related papers
- Enhancing Underwater Imaging with 4-D Light Fields: Dataset and Method [77.80712860663886]
4-D light fields (LFs) enhance underwater imaging plagued by light absorption, scattering, and other challenges.
We propose a progressive framework for underwater 4-D LF image enhancement and depth estimation.
We construct the first 4-D LF-based underwater image dataset for quantitative evaluation and supervised training of learning-based methods.
arXiv Detail & Related papers (2024-08-30T15:06:45Z) - Towards Ultra-High-Definition Image Deraining: A Benchmark and An Efficient Method [42.331058889312466]
This paper contributes the first large-scale UHD image deraining dataset, 4K-Rain13k, that contains 13,000 image pairs at 4K resolution.
We develop an effective and efficient vision-based architecture (UDR-Mixer) to better solve this task.
arXiv Detail & Related papers (2024-05-27T11:45:08Z) - AADNet: Attention aware Demoiréing Network [2.1626093085892144]
Moire pattern frequently appears in photographs captured with mobile devices and digital cameras.
We propose a novel lightweight architecture, AADNet, for high-resolution image demoire'ing.
arXiv Detail & Related papers (2024-03-13T09:48:11Z) - Recurrent Multi-scale Transformer for High-Resolution Salient Object
Detection [68.65338791283298]
Salient Object Detection (SOD) aims to identify and segment the most conspicuous objects in an image or video.
Traditional SOD methods are largely limited to low-resolution images, making them difficult to adapt to the development of High-Resolution SOD.
In this work, we first propose a new HRS10K dataset, which contains 10,500 high-quality annotated images at 2K-8K resolution.
arXiv Detail & Related papers (2023-08-07T17:49:04Z) - 4K-HAZE: A Dehazing Benchmark with 4K Resolution Hazy and Haze-Free
Images [12.402054374952485]
We develop a novel method to simulate 4K hazy images from clear images, which first estimates the scene depth, simulates the light rays and object reflectance, then migrates the synthetic images to real domains by using a GAN.
We wrap these synthesized images into a benchmark called the 4K-HAZE dataset.
The most appealing aspect of our approach is the capability to run a 4K image on a single GPU with 24G RAM in real-time (33fps)
arXiv Detail & Related papers (2023-03-28T09:39:29Z) - Ultra-High-Definition Low-Light Image Enhancement: A Benchmark and
Transformer-Based Method [51.30748775681917]
We consider the task of low-light image enhancement (LLIE) and introduce a large-scale database consisting of images at 4K and 8K resolution.
We conduct systematic benchmarking studies and provide a comparison of current LLIE algorithms.
As a second contribution, we introduce LLFormer, a transformer-based low-light enhancement method.
arXiv Detail & Related papers (2022-12-22T09:05:07Z) - Perceptual Image Enhancement for Smartphone Real-Time Applications [60.45737626529091]
We propose LPIENet, a lightweight network for perceptual image enhancement.
Our model can deal with noise artifacts, diffraction artifacts, blur, and HDR overexposure.
Our model can process 2K resolution images under 1 second in mid-level commercial smartphones.
arXiv Detail & Related papers (2022-10-24T19:16:33Z) - Highly Accurate Dichotomous Image Segmentation [139.79513044546]
A new task called dichotomous image segmentation (DIS) aims to segment highly accurate objects from natural images.
We collect the first large-scale dataset, DIS5K, which contains 5,470 high-resolution (e.g., 2K, 4K or larger) images.
We also introduce a simple intermediate supervision baseline (IS-Net) using both feature-level and mask-level guidance for DIS model training.
arXiv Detail & Related papers (2022-03-06T20:09:19Z) - ORStereo: Occlusion-Aware Recurrent Stereo Matching for 4K-Resolution
Images [13.508624751092654]
We present the Occlusion-aware Recurrent binocular Stereo matching (ORStereo)
ORStereo generalizes to unseen high-resolution images with large disparity ranges by formulating the task as residual updates and refinements of an initial prediction.
We test the model's capability on both synthetic and real-world high-resolution images.
arXiv Detail & Related papers (2021-03-13T21:46:06Z) - Contextual Residual Aggregation for Ultra High-Resolution Image
Inpainting [12.839962012888199]
We propose a Contextual Residual Aggregation (CRA) mechanism that can produce high-frequency residuals for missing contents.
CRA mechanism produces high-frequency residuals for missing contents by weighted aggregating residuals from contextual patches.
We train the proposed model on small images with resolutions 512x512 and perform inference on high-resolution images, achieving compelling inpainting quality.
arXiv Detail & Related papers (2020-05-19T18:55:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.