Implicit Transformer Network for Screen Content Image Continuous
Super-Resolution
- URL: http://arxiv.org/abs/2112.06174v1
- Date: Sun, 12 Dec 2021 07:39:37 GMT
- Title: Implicit Transformer Network for Screen Content Image Continuous
Super-Resolution
- Authors: Jingyu Yang, Sheng Shen, Huanjing Yue, Kun Li
- Abstract summary: High-resolution (HR) screen contents may be downsampled and compressed.
Super-resolution (SR) of low-resolution (LR) screen content images (SCIs) is highly demanded by the HR display or by the users to zoom in for detail observation.
We propose a novel Implicit Transformer Super-Resolution Network (ITSRN) for SCISR.
- Score: 27.28782217250359
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Nowadays, there is an explosive growth of screen contents due to the wide
application of screen sharing, remote cooperation, and online education. To
match the limited terminal bandwidth, high-resolution (HR) screen contents may
be downsampled and compressed. At the receiver side, the super-resolution (SR)
of low-resolution (LR) screen content images (SCIs) is highly demanded by the
HR display or by the users to zoom in for detail observation. However, image SR
methods mostly designed for natural images do not generalize well for SCIs due
to the very different image characteristics as well as the requirement of SCI
browsing at arbitrary scales. To this end, we propose a novel Implicit
Transformer Super-Resolution Network (ITSRN) for SCISR. For high-quality
continuous SR at arbitrary ratios, pixel values at query coordinates are
inferred from image features at key coordinates by the proposed implicit
transformer and an implicit position encoding scheme is proposed to aggregate
similar neighboring pixel values to the query one. We construct benchmark SCI1K
and SCI1K-compression datasets with LR and HR SCI pairs. Extensive experiments
show that the proposed ITSRN significantly outperforms several competitive
continuous and discrete SR methods for both compressed and uncompressed SCIs.
Related papers
- OPE-SR: Orthogonal Position Encoding for Designing a Parameter-free
Upsampling Module in Arbitrary-scale Image Super-Resolution [11.74426147465809]
Implicit neural representation (INR) is a popular approach for arbitrary-scale image super-resolution.
We propose an OPE-Upscale module to replace the INR-based upsampling module for arbitrary-scale image super-resolution.
arXiv Detail & Related papers (2023-03-02T09:26:14Z) - Reference-based Image and Video Super-Resolution via C2-Matching [100.0808130445653]
We propose C2-Matching, which performs explicit robust matching crossing transformation and resolution.
C2-Matching significantly outperforms state of the arts on the standard CUFED5 benchmark.
We also extend C2-Matching to Reference-based Video Super-Resolution task, where an image taken in a similar scene serves as the HR reference image.
arXiv Detail & Related papers (2022-12-19T16:15:02Z) - Contextual Learning in Fourier Complex Field for VHR Remote Sensing
Images [64.84260544255477]
transformer-based models demonstrated outstanding potential for learning high-order contextual relationships from natural images with general resolution (224x224 pixels)
We propose a complex self-attention (CSA) mechanism to model the high-order contextual information with less than half computations of naive SA.
By stacking various layers of CSA blocks, we propose the Fourier Complex Transformer (FCT) model to learn global contextual information from VHR aerial images.
arXiv Detail & Related papers (2022-10-28T08:13:33Z) - ITSRN++: Stronger and Better Implicit Transformer Network for Continuous
Screen Content Image Super-Resolution [32.441761727608856]
The proposed method achieves state-of-the-art performance for SCI SR (outperforming SwinIR by 0.74 dB for x3 SR) and also works well for natural image SR.
We construct a large scale SCI2K dataset to facilitate the research on SCI SR.
arXiv Detail & Related papers (2022-10-17T07:47:34Z) - CADyQ: Content-Aware Dynamic Quantization for Image Super-Resolution [55.50793823060282]
We propose a novel Content-Aware Dynamic Quantization (CADyQ) method for image super-resolution (SR) networks.
CADyQ allocates optimal bits to local regions and layers adaptively based on the local contents of an input image.
The pipeline has been tested on various SR networks and evaluated on several standard benchmarks.
arXiv Detail & Related papers (2022-07-21T07:50:50Z) - SRWarp: Generalized Image Super-Resolution under Arbitrary
Transformation [65.88321755969677]
Deep CNNs have achieved significant successes in image processing and its applications, including single image super-resolution.
Recent approaches extend the scope to real-valued upsampling factors.
We propose the SRWarp framework to further generalize the SR tasks toward an arbitrary image transformation.
arXiv Detail & Related papers (2021-04-21T02:50:41Z) - Hyperspectral Image Super-resolution via Deep Progressive Zero-centric
Residual Learning [62.52242684874278]
Cross-modality distribution of spatial and spectral information makes the problem challenging.
We propose a novel textitlightweight deep neural network-based framework, namely PZRes-Net.
Our framework learns a high resolution and textitzero-centric residual image, which contains high-frequency spatial details of the scene.
arXiv Detail & Related papers (2020-06-18T06:32:11Z) - DDet: Dual-path Dynamic Enhancement Network for Real-World Image
Super-Resolution [69.2432352477966]
Real image super-resolution(Real-SR) focus on the relationship between real-world high-resolution(HR) and low-resolution(LR) image.
In this article, we propose a Dual-path Dynamic Enhancement Network(DDet) for Real-SR.
Unlike conventional methods which stack up massive convolutional blocks for feature representation, we introduce a content-aware framework to study non-inherently aligned image pair.
arXiv Detail & Related papers (2020-02-25T18:24:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.