Adaptive Local Implicit Image Function for Arbitrary-scale
Super-resolution
- URL: http://arxiv.org/abs/2208.04318v1
- Date: Sun, 7 Aug 2022 11:23:23 GMT
- Title: Adaptive Local Implicit Image Function for Arbitrary-scale
Super-resolution
- Authors: Hongwei Li, Tao Dai, Yiming Li, Xueyi Zou, Shu-Tao Xia
- Abstract summary: Local implicit image function (LIIF) denotes images as a continuous function where pixel values are expansion by using the corresponding coordinates as inputs.
LIIF can be adopted for arbitrary-scale image super-resolution tasks, resulting in a single effective and efficient model for various up-scaling factors.
We propose a novel adaptive local image function (A-LIIF) to alleviate this problem.
- Score: 61.95533972380704
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Image representation is critical for many visual tasks. Instead of
representing images discretely with 2D arrays of pixels, a recent study, namely
local implicit image function (LIIF), denotes images as a continuous function
where pixel values are expansion by using the corresponding coordinates as
inputs. Due to its continuous nature, LIIF can be adopted for arbitrary-scale
image super-resolution tasks, resulting in a single effective and efficient
model for various up-scaling factors. However, LIIF often suffers from
structural distortions and ringing artifacts around edges, mostly because all
pixels share the same model, thus ignoring the local properties of the image.
In this paper, we propose a novel adaptive local image function (A-LIIF) to
alleviate this problem. Specifically, our A-LIIF consists of two main
components: an encoder and a expansion network. The former captures cross-scale
image features, while the latter models the continuous up-scaling function by a
weighted combination of multiple local implicit image functions. Accordingly,
our A-LIIF can reconstruct the high-frequency textures and structures more
accurately. Experiments on multiple benchmark datasets verify the effectiveness
of our method. Our codes are available at
\url{https://github.com/LeeHW-THU/A-LIIF}.
Related papers
- Image-GS: Content-Adaptive Image Representation via 2D Gaussians [55.15950594752051]
We propose Image-GS, a content-adaptive image representation.
Using anisotropic 2D Gaussians as the basis, Image-GS shows high memory efficiency, supports fast random access, and offers a natural level of detail stack.
General efficiency and fidelity of Image-GS are validated against several recent neural image representations and industry-standard texture compressors.
We hope this research offers insights for developing new applications that require adaptive quality and resource control, such as machine perception, asset streaming, and content generation.
arXiv Detail & Related papers (2024-07-02T00:45:21Z) - CricaVPR: Cross-image Correlation-aware Representation Learning for Visual Place Recognition [73.51329037954866]
We propose a robust global representation method with cross-image correlation awareness for visual place recognition.
Our method uses the attention mechanism to correlate multiple images within a batch.
Our method outperforms state-of-the-art methods by a large margin with significantly less training time.
arXiv Detail & Related papers (2024-02-29T15:05:11Z) - Dynamic Implicit Image Function for Efficient Arbitrary-Scale Image
Representation [24.429100808481394]
We propose Dynamic Implicit Image Function (DIIF), which is a fast and efficient method to represent images with arbitrary resolution.
We propose a coordinate grouping and slicing strategy, which enables the neural network to perform decoding from coordinate slices to pixel value slices.
With dynamic coordinate slicing, DIIF significantly reduces the computational cost when encountering arbitrary-scale SR.
arXiv Detail & Related papers (2023-06-21T15:04:34Z) - T-former: An Efficient Transformer for Image Inpainting [50.43302925662507]
A class of attention-based network architectures, called transformer, has shown significant performance on natural language processing fields.
In this paper, we design a novel attention linearly related to the resolution according to Taylor expansion, and based on this attention, a network called $T$-former is designed for image inpainting.
Experiments on several benchmark datasets demonstrate that our proposed method achieves state-of-the-art accuracy while maintaining a relatively low number of parameters and computational complexity.
arXiv Detail & Related papers (2023-05-12T04:10:42Z) - Single Image Super-Resolution via a Dual Interactive Implicit Neural
Network [5.331665215168209]
We introduce a novel implicit neural network for the task of single image super-resolution at arbitrary scale factors.
We demonstrate the efficacy and flexibility of our approach against the state of the art on publicly available benchmark datasets.
arXiv Detail & Related papers (2022-10-23T02:05:19Z) - InfinityGAN: Towards Infinite-Resolution Image Synthesis [92.40782797030977]
We present InfinityGAN, a method to generate arbitrary-resolution images.
We show how it trains and infers patch-by-patch seamlessly with low computational resources.
arXiv Detail & Related papers (2021-04-08T17:59:30Z) - Learning Continuous Image Representation with Local Implicit Image
Function [21.27344998709831]
We propose LIIF representation, which takes an image coordinate and the 2D deep features around the coordinate as inputs, predicts the RGB value at a given coordinate as an output.
To generate the continuous representation for images, we train an encoder with LIIF representation via a self-supervised task with super-resolution.
The learned continuous representation can be presented in arbitrary resolution even extrapolate to x30 higher resolution.
arXiv Detail & Related papers (2020-12-16T18:56:50Z) - Adversarial Generation of Continuous Images [31.92891885615843]
In this paper, we propose two novel architectural techniques for building INR-based image decoders.
We use them to build a state-of-the-art continuous image GAN.
Our proposed INR-GAN architecture improves the performance of continuous image generators by several times.
arXiv Detail & Related papers (2020-11-24T11:06:40Z) - Funnel Activation for Visual Recognition [92.18474421444377]
We present a conceptually simple but effective funnel activation for image recognition tasks, called Funnel activation (FReLU)
FReLU extends ReLU and PReLU to a 2D activation by adding a negligible overhead of spatial condition.
We conduct experiments on ImageNet, COCO detection, and semantic segmentation, showing great improvements and robustness of FReLU in the visual recognition tasks.
arXiv Detail & Related papers (2020-07-23T07:02:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.