A-ESRGAN: Training Real-World Blind Super-Resolution with Attention
U-Net Discriminators
- URL: http://arxiv.org/abs/2112.10046v1
- Date: Sun, 19 Dec 2021 02:50:23 GMT
- Title: A-ESRGAN: Training Real-World Blind Super-Resolution with Attention
U-Net Discriminators
- Authors: Zihao Wei, Yidong Huang, Yuang Chen, Chenhao Zheng, Jinnan Gao
- Abstract summary: Blind image super-resolution(SR) is a long-standing task in CV that aims to restore low-resolution images suffering from unknown and complex distortions.
We present A-ESRGAN, a GAN model for blind SR tasks featuring an attention U-Net based, multi-scale discriminator.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Blind image super-resolution(SR) is a long-standing task in CV that aims to
restore low-resolution images suffering from unknown and complex distortions.
Recent work has largely focused on adopting more complicated degradation models
to emulate real-world degradations. The resulting models have made
breakthroughs in perceptual loss and yield perceptually convincing results.
However, the limitation brought by current generative adversarial network
structures is still significant: treating pixels equally leads to the ignorance
of the image's structural features, and results in performance drawbacks such
as twisted lines and background over-sharpening or blurring. In this paper, we
present A-ESRGAN, a GAN model for blind SR tasks featuring an attention U-Net
based, multi-scale discriminator that can be seamlessly integrated with other
generators. To our knowledge, this is the first work to introduce attention
U-Net structure as the discriminator of GAN to solve blind SR problems. And the
paper also gives an interpretation for the mechanism behind multi-scale
attention U-Net that brings performance breakthrough to the model. Through
comparison experiments with prior works, our model presents state-of-the-art
level performance on the non-reference natural image quality evaluator metric.
And our ablation studies have shown that with our discriminator, the RRDB based
generator can leverage the structural features of an image in multiple scales,
and consequently yields more perceptually realistic high-resolution images
compared to prior works.
Related papers
- RBSR: Efficient and Flexible Recurrent Network for Burst
Super-Resolution [57.98314517861539]
Burst super-resolution (BurstSR) aims at reconstructing a high-resolution (HR) image from a sequence of low-resolution (LR) and noisy images.
In this paper, we suggest fusing cues frame-by-frame with an efficient and flexible recurrent network.
arXiv Detail & Related papers (2023-06-30T12:14:13Z) - TcGAN: Semantic-Aware and Structure-Preserved GANs with Individual
Vision Transformer for Fast Arbitrary One-Shot Image Generation [11.207512995742999]
One-shot image generation (OSG) with generative adversarial networks that learn from the internal patches of a given image has attracted world wide attention.
We propose a novel structure-preserved method TcGAN with individual vision transformer to overcome the shortcomings of the existing one-shot image generation methods.
arXiv Detail & Related papers (2023-02-16T03:05:59Z) - Hierarchical Similarity Learning for Aliasing Suppression Image
Super-Resolution [64.15915577164894]
A hierarchical image super-resolution network (HSRNet) is proposed to suppress the influence of aliasing.
HSRNet achieves better quantitative and visual performance than other works, and remits the aliasing more effectively.
arXiv Detail & Related papers (2022-06-07T14:55:32Z) - Memory-augmented Deep Unfolding Network for Guided Image
Super-resolution [67.83489239124557]
Guided image super-resolution (GISR) aims to obtain a high-resolution (HR) target image by enhancing the spatial resolution of a low-resolution (LR) target image under the guidance of a HR image.
Previous model-based methods mainly takes the entire image as a whole, and assume the prior distribution between the HR target image and the HR guidance image.
We propose a maximal a posterior (MAP) estimation model for GISR with two types of prior on the HR target image.
arXiv Detail & Related papers (2022-02-12T15:37:13Z) - Best-Buddy GANs for Highly Detailed Image Super-Resolution [71.13466303340192]
We consider the single image super-resolution (SISR) problem, where a high-resolution (HR) image is generated based on a low-resolution (LR) input.
Most methods along this line rely on a predefined single-LR-single-HR mapping, which is not flexible enough for the SISR task.
We propose best-buddy GANs (Beby-GAN) for rich-detail SISR. Relaxing the immutable one-to-one constraint, we allow the estimated patches to dynamically seek the best supervision.
arXiv Detail & Related papers (2021-03-29T02:58:27Z) - Learning Structral coherence Via Generative Adversarial Network for
Single Image Super-Resolution [13.803141755183827]
Recent generative adversarial network (GAN) based SISR methods have yielded overall realistic SR images.
We introduce the gradient branch into the generator to preserve structural information by restoring high-resolution gradient maps in SR process.
In addition, we utilize a U-net based discriminator to consider both the whole image and the detailed per-pixel authenticity.
arXiv Detail & Related papers (2021-01-25T15:26:23Z) - Iterative Network for Image Super-Resolution [69.07361550998318]
Single image super-resolution (SISR) has been greatly revitalized by the recent development of convolutional neural networks (CNN)
This paper provides a new insight on conventional SISR algorithm, and proposes a substantially different approach relying on the iterative optimization.
A novel iterative super-resolution network (ISRN) is proposed on top of the iterative optimization.
arXiv Detail & Related papers (2020-05-20T11:11:47Z) - Enhancing Perceptual Loss with Adversarial Feature Matching for
Super-Resolution [5.258555266148511]
Single image super-resolution (SISR) is an ill-posed problem with an indeterminate number of valid solutions.
We show that the root cause of these pattern artifacts can be traced back to a mismatch between the pre-training objective of perceptual loss and the super-resolved objective.
arXiv Detail & Related papers (2020-05-15T12:36:54Z) - Deep Generative Adversarial Residual Convolutional Networks for
Real-World Super-Resolution [31.934084942626257]
We propose a deep Super-Resolution Residual Convolutional Generative Adversarial Network (SRResCGAN)
It follows the real-world degradation settings by adversarial training the model with pixel-wise supervision in the HR domain from its generated LR counterpart.
The proposed network exploits the residual learning by minimizing the energy-based objective function with powerful image regularization and convex optimization techniques.
arXiv Detail & Related papers (2020-05-03T00:12:38Z) - A U-Net Based Discriminator for Generative Adversarial Networks [86.67102929147592]
We propose an alternative U-Net based discriminator architecture for generative adversarial networks (GANs)
The proposed architecture allows to provide detailed per-pixel feedback to the generator while maintaining the global coherence of synthesized images.
The novel discriminator improves over the state of the art in terms of the standard distribution and image quality metrics.
arXiv Detail & Related papers (2020-02-28T11:16:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.