Enhancing Multi-Scale Implicit Learning in Image Super-Resolution with
Integrated Positional Encoding
- URL: http://arxiv.org/abs/2112.05756v1
- Date: Fri, 10 Dec 2021 06:09:55 GMT
- Title: Enhancing Multi-Scale Implicit Learning in Image Super-Resolution with
Integrated Positional Encoding
- Authors: Ying-Tian Liu, Yuan-Chen Guo, Song-Hai Zhang
- Abstract summary: We consider each pixel as the aggregation of signals from a local area in an image super-resolution context.
We propose integrated positional encoding (IPE), extending traditional positional encoding by aggregating frequency information over the pixel area.
We show the effectiveness of IPE-LIIF by quantitative and qualitative evaluations, and further demonstrate the generalization ability of IPE to larger image scales.
- Score: 4.781615891172263
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Is the center position fully capable of representing a pixel? There is
nothing wrong to represent pixels with their centers in a discrete image
representation, but it makes more sense to consider each pixel as the
aggregation of signals from a local area in an image super-resolution (SR)
context. Despite the great capability of coordinate-based implicit
representation in the field of arbitrary-scale image SR, this area's nature of
pixels is not fully considered. To this end, we propose integrated positional
encoding (IPE), extending traditional positional encoding by aggregating
frequency information over the pixel area. We apply IPE to the state-of-the-art
arbitrary-scale image super-resolution method: local implicit image function
(LIIF), presenting IPE-LIIF. We show the effectiveness of IPE-LIIF by
quantitative and qualitative evaluations, and further demonstrate the
generalization ability of IPE to larger image scales and multiple
implicit-based methods. Code will be released.
Related papers
- Exploring Multi-view Pixel Contrast for General and Robust Image Forgery Localization [4.8454936010479335]
We propose a Multi-view Pixel-wise Contrastive algorithm (MPC) for image forgery localization.
Specifically, we first pre-train the backbone network with the supervised contrastive loss.
Then the localization head is fine-tuned using the cross-entropy loss, resulting in a better pixel localizer.
arXiv Detail & Related papers (2024-06-19T13:51:52Z) - Pixel-Inconsistency Modeling for Image Manipulation Localization [63.54342601757723]
Digital image forensics plays a crucial role in image authentication and manipulation localization.
This paper presents a generalized and robust manipulation localization model through the analysis of pixel inconsistency artifacts.
Experiments show that our method successfully extracts inherent pixel-inconsistency forgery fingerprints.
arXiv Detail & Related papers (2023-09-30T02:54:51Z) - Dynamic Implicit Image Function for Efficient Arbitrary-Scale Image
Representation [24.429100808481394]
We propose Dynamic Implicit Image Function (DIIF), which is a fast and efficient method to represent images with arbitrary resolution.
We propose a coordinate grouping and slicing strategy, which enables the neural network to perform decoding from coordinate slices to pixel value slices.
With dynamic coordinate slicing, DIIF significantly reduces the computational cost when encountering arbitrary-scale SR.
arXiv Detail & Related papers (2023-06-21T15:04:34Z) - OPE-SR: Orthogonal Position Encoding for Designing a Parameter-free
Upsampling Module in Arbitrary-scale Image Super-Resolution [11.74426147465809]
Implicit neural representation (INR) is a popular approach for arbitrary-scale image super-resolution.
We propose an OPE-Upscale module to replace the INR-based upsampling module for arbitrary-scale image super-resolution.
arXiv Detail & Related papers (2023-03-02T09:26:14Z) - Single Image Super-Resolution via a Dual Interactive Implicit Neural
Network [5.331665215168209]
We introduce a novel implicit neural network for the task of single image super-resolution at arbitrary scale factors.
We demonstrate the efficacy and flexibility of our approach against the state of the art on publicly available benchmark datasets.
arXiv Detail & Related papers (2022-10-23T02:05:19Z) - Adaptive Local Implicit Image Function for Arbitrary-scale
Super-resolution [61.95533972380704]
Local implicit image function (LIIF) denotes images as a continuous function where pixel values are expansion by using the corresponding coordinates as inputs.
LIIF can be adopted for arbitrary-scale image super-resolution tasks, resulting in a single effective and efficient model for various up-scaling factors.
We propose a novel adaptive local image function (A-LIIF) to alleviate this problem.
arXiv Detail & Related papers (2022-08-07T11:23:23Z) - Deep Posterior Distribution-based Embedding for Hyperspectral Image
Super-resolution [75.24345439401166]
This paper focuses on how to embed the high-dimensional spatial-spectral information of hyperspectral (HS) images efficiently and effectively.
We formulate HS embedding as an approximation of the posterior distribution of a set of carefully-defined HS embedding events.
Then, we incorporate the proposed feature embedding scheme into a source-consistent super-resolution framework that is physically-interpretable.
Experiments over three common benchmark datasets demonstrate that PDE-Net achieves superior performance over state-of-the-art methods.
arXiv Detail & Related papers (2022-05-30T06:59:01Z) - Learning Enriched Features for Fast Image Restoration and Enhancement [166.17296369600774]
This paper presents a holistic goal of maintaining spatially-precise high-resolution representations through the entire network.
We learn an enriched set of features that combines contextual information from multiple scales, while simultaneously preserving the high-resolution spatial details.
Our approach achieves state-of-the-art results for a variety of image processing tasks, including defocus deblurring, image denoising, super-resolution, and image enhancement.
arXiv Detail & Related papers (2022-04-19T17:59:45Z) - Steering Self-Supervised Feature Learning Beyond Local Pixel Statistics [60.92229707497999]
We introduce a novel principle for self-supervised feature learning based on the discrimination of specific transformations of an image.
We demonstrate experimentally that learning to discriminate transformations such as LCI, image warping and rotations, yields features with state of the art generalization capabilities.
arXiv Detail & Related papers (2020-04-05T22:09:08Z) - A U-Net Based Discriminator for Generative Adversarial Networks [86.67102929147592]
We propose an alternative U-Net based discriminator architecture for generative adversarial networks (GANs)
The proposed architecture allows to provide detailed per-pixel feedback to the generator while maintaining the global coherence of synthesized images.
The novel discriminator improves over the state of the art in terms of the standard distribution and image quality metrics.
arXiv Detail & Related papers (2020-02-28T11:16:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.