SRTransGAN: Image Super-Resolution using Transformer based Generative
Adversarial Network
- URL: http://arxiv.org/abs/2312.01999v1
- Date: Mon, 4 Dec 2023 16:22:39 GMT
- Title: SRTransGAN: Image Super-Resolution using Transformer based Generative
Adversarial Network
- Authors: Neeraj Baghel, Shiv Ram Dubey, Satish Kumar Singh
- Abstract summary: We propose a transformer-based encoder-decoder network as a generator to generate 2x images and 4x images.
The proposed SRTransGAN outperforms the existing methods by 4.38 % on an average of PSNR and SSIM scores.
- Score: 16.243363392717434
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Image super-resolution aims to synthesize high-resolution image from a
low-resolution image. It is an active area to overcome the resolution
limitations in several applications like low-resolution object-recognition,
medical image enhancement, etc. The generative adversarial network (GAN) based
methods have been the state-of-the-art for image super-resolution by utilizing
the convolutional neural networks (CNNs) based generator and discriminator
networks. However, the CNNs are not able to exploit the global information very
effectively in contrast to the transformers, which are the recent breakthrough
in deep learning by exploiting the self-attention mechanism. Motivated from the
success of transformers in language and vision applications, we propose a
SRTransGAN for image super-resolution using transformer based GAN.
Specifically, we propose a novel transformer-based encoder-decoder network as a
generator to generate 2x images and 4x images. We design the discriminator
network using vision transformer which uses the image as sequence of patches
and hence useful for binary classification between synthesized and real
high-resolution images. The proposed SRTransGAN outperforms the existing
methods by 4.38 % on an average of PSNR and SSIM scores. We also analyze the
saliency map to understand the learning ability of the proposed method.
Related papers
- CWT-Net: Super-resolution of Histopathology Images Using a Cross-scale Wavelet-based Transformer [15.930878163092983]
Super-resolution (SR) aims to enhance the quality of low-resolution images and has been widely applied in medical imaging.
We propose a novel network called CWT-Net, which leverages cross-scale image wavelet transform and Transformer architecture.
Our model significantly outperforms state-of-the-art methods in both performance and visualization evaluations.
arXiv Detail & Related papers (2024-09-11T08:26:28Z) - PTSR: Patch Translator for Image Super-Resolution [16.243363392717434]
We propose a patch translator for image super-resolution (PTSR) to address this problem.
The proposed PTSR is a transformer-based GAN network with no convolution operation.
We introduce a novel patch translator module for regenerating the improved patches utilising multi-head attention.
arXiv Detail & Related papers (2023-10-20T01:45:00Z) - Transformer-based Generative Adversarial Networks in Computer Vision: A
Comprehensive Survey [26.114550071165628]
Generative Adversarial Networks (GANs) have been very successful for synthesizing the images in a given dataset.
Recent works have tried to exploit the Transformers in GAN framework for the image/video synthesis.
This paper presents a comprehensive survey on the developments and advancements in GANs utilizing the Transformer networks for computer vision applications.
arXiv Detail & Related papers (2023-02-17T01:13:58Z) - Cross-receptive Focused Inference Network for Lightweight Image
Super-Resolution [64.25751738088015]
Transformer-based methods have shown impressive performance in single image super-resolution (SISR) tasks.
Transformers that need to incorporate contextual information to extract features dynamically are neglected.
We propose a lightweight Cross-receptive Focused Inference Network (CFIN) that consists of a cascade of CT Blocks mixed with CNN and Transformer.
arXiv Detail & Related papers (2022-07-06T16:32:29Z) - Rich CNN-Transformer Feature Aggregation Networks for Super-Resolution [50.10987776141901]
Recent vision transformers along with self-attention have achieved promising results on various computer vision tasks.
We introduce an effective hybrid architecture for super-resolution (SR) tasks, which leverages local features from CNNs and long-range dependencies captured by transformers.
Our proposed method achieves state-of-the-art SR results on numerous benchmark datasets.
arXiv Detail & Related papers (2022-03-15T06:52:25Z) - Transformer-based SAR Image Despeckling [53.99620005035804]
We introduce a transformer-based network for SAR image despeckling.
The proposed despeckling network comprises of a transformer-based encoder which allows the network to learn global dependencies between different image regions.
Experiments show that the proposed method achieves significant improvements over traditional and convolutional neural network-based despeckling methods.
arXiv Detail & Related papers (2022-01-23T20:09:01Z) - CSformer: Bridging Convolution and Transformer for Compressive Sensing [65.22377493627687]
This paper proposes a hybrid framework that integrates the advantages of leveraging detailed spatial information from CNN and the global context provided by transformer for enhanced representation learning.
The proposed approach is an end-to-end compressive image sensing method, composed of adaptive sampling and recovery.
The experimental results demonstrate the effectiveness of the dedicated transformer-based architecture for compressive sensing.
arXiv Detail & Related papers (2021-12-31T04:37:11Z) - The Nuts and Bolts of Adopting Transformer in GANs [124.30856952272913]
We investigate the properties of Transformer in the generative adversarial network (GAN) framework for high-fidelity image synthesis.
Our study leads to a new alternative design of Transformers in GAN, a convolutional neural network (CNN)-free generator termed as STrans-G.
arXiv Detail & Related papers (2021-10-25T17:01:29Z) - Fusformer: A Transformer-based Fusion Approach for Hyperspectral Image
Super-resolution [9.022005574190182]
We design a network based on the transformer for fusing the low-resolution hyperspectral images and high-resolution multispectral images.
Considering the LR-HSIs hold the main spectral structure, the network focuses on the spatial detail estimation.
Various experiments and quality indexes show our approach's superiority compared with other state-of-the-art methods.
arXiv Detail & Related papers (2021-09-05T14:00:34Z) - Learning Enriched Features for Real Image Restoration and Enhancement [166.17296369600774]
convolutional neural networks (CNNs) have achieved dramatic improvements over conventional approaches for image restoration task.
We present a novel architecture with the collective goals of maintaining spatially-precise high-resolution representations through the entire network.
Our approach learns an enriched set of features that combines contextual information from multiple scales, while simultaneously preserving the high-resolution spatial details.
arXiv Detail & Related papers (2020-03-15T11:04:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.