UltraSR: Spatial Encoding is a Missing Key for Implicit Image
Function-based Arbitrary-Scale Super-Resolution
- URL: http://arxiv.org/abs/2103.12716v1
- Date: Tue, 23 Mar 2021 17:36:42 GMT
- Title: UltraSR: Spatial Encoding is a Missing Key for Implicit Image
Function-based Arbitrary-Scale Super-Resolution
- Authors: Xingqian Xu, Zhangyang Wang, Humphrey Shi
- Abstract summary: In this work, we propose UltraSR, a simple yet effective new network design based on implicit image functions.
We show that spatial encoding is indeed a missing key towards the next-stage high-accuracy implicit image function.
Our UltraSR sets new state-of-the-art performance on the DIV2K benchmark under all super-resolution scales.
- Score: 74.82282301089994
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: The recent success of NeRF and other related implicit neural representation
methods has opened a new path for continuous image representation, where pixel
values no longer need to be looked up from stored discrete 2D arrays but can be
inferred from neural network models on a continuous spatial domain. Although
the recent work LIIF has demonstrated that such novel approach can achieve good
performance on the arbitrary-scale super-resolution task, their upscaled images
frequently show structural distortion due to the faulty prediction on
high-frequency textures. In this work, we propose UltraSR, a simple yet
effective new network design based on implicit image functions in which spatial
coordinates and periodic encoding are deeply integrated with the implicit
neural representation. We show that spatial encoding is indeed a missing key
towards the next-stage high-accuracy implicit image function through extensive
experiments and ablation studies. Our UltraSR sets new state-of-the-art
performance on the DIV2K benchmark under all super-resolution scales comparing
to previous state-of-the-art methods. UltraSR also achieves superior
performance on other standard benchmark datasets in which it outperforms prior
works in almost all experiments. Our code will be released at
https://github.com/SHI-Labs/UltraSR-Arbitrary-Scale-Super-Resolution.
Related papers
- HF-Diff: High-Frequency Perceptual Loss and Distribution Matching for One-Step Diffusion-Based Image Super-Resolution [0.0]
We introduce a high-frequency perceptual loss by utilizing an invertible neural network (INN) pretrained on the ImageNet dataset.
During the training phase, we impose to preserve the high-frequency features of super-resolved and ground truth (GT) images.
We achieve a state-of-the-art CLIPIQA score in the benchmark RealSR, RealSet65, DIV2K-Val, and ImageNet datasets.
arXiv Detail & Related papers (2024-11-20T18:56:24Z) - Super-Resolution Neural Operator [5.018040244860608]
We propose a framework that can resolve high-resolution (HR) images at arbitrary scales from the low-resolution (LR) counterparts.
Treating the LR-HR image pairs as continuous functions approximated with different grid sizes, SRNO learns the mapping between the corresponding function spaces.
Experiments show that SRNO outperforms existing continuous SR methods in terms of both accuracy and running time.
arXiv Detail & Related papers (2023-03-05T06:17:43Z) - Hybrid Pixel-Unshuffled Network for Lightweight Image Super-Resolution [64.54162195322246]
Convolutional neural network (CNN) has achieved great success on image super-resolution (SR)
Most deep CNN-based SR models take massive computations to obtain high performance.
We propose a novel Hybrid Pixel-Unshuffled Network (HPUN) by introducing an efficient and effective downsampling module into the SR task.
arXiv Detail & Related papers (2022-03-16T20:10:41Z) - NeRF-SR: High-Quality Neural Radiance Fields using Super-Sampling [82.99453001445478]
We present NeRF-SR, a solution for high-resolution (HR) novel view synthesis with mostly low-resolution (LR) inputs.
Our method is built upon Neural Radiance Fields (NeRF) that predicts per-point density and color with a multi-layer perceptron.
arXiv Detail & Related papers (2021-12-03T07:33:47Z) - Local Texture Estimator for Implicit Representation Function [10.165529175855712]
Local Texture Estimator (LTE) is a dominant-frequency estimator for natural images.
LTE is capable of characterizing image textures in 2D Fourier space.
We show that an LTE-based neural function outperforms existing deep SR methods within an arbitrary-scale.
arXiv Detail & Related papers (2021-11-17T06:01:17Z) - LAPAR: Linearly-Assembled Pixel-Adaptive Regression Network for Single
Image Super-Resolution and Beyond [75.37541439447314]
Single image super-resolution (SISR) deals with a fundamental problem of upsampling a low-resolution (LR) image to its high-resolution (HR) version.
This paper proposes a linearly-assembled pixel-adaptive regression network (LAPAR) to strike a sweet spot of deep model complexity and resulting SISR quality.
arXiv Detail & Related papers (2021-05-21T15:47:18Z) - Asymmetric CNN for image super-resolution [102.96131810686231]
Deep convolutional neural networks (CNNs) have been widely applied for low-level vision over the past five years.
We propose an asymmetric CNN (ACNet) comprising an asymmetric block (AB), a mem?ory enhancement block (MEB) and a high-frequency feature enhancement block (HFFEB) for image super-resolution.
Our ACNet can effectively address single image super-resolution (SISR), blind SISR and blind SISR of blind noise problems.
arXiv Detail & Related papers (2021-03-25T07:10:46Z) - Hyperspectral Image Super-resolution via Deep Progressive Zero-centric
Residual Learning [62.52242684874278]
Cross-modality distribution of spatial and spectral information makes the problem challenging.
We propose a novel textitlightweight deep neural network-based framework, namely PZRes-Net.
Our framework learns a high resolution and textitzero-centric residual image, which contains high-frequency spatial details of the scene.
arXiv Detail & Related papers (2020-06-18T06:32:11Z) - Generalized Octave Convolutions for Learned Multi-Frequency Image
Compression [20.504561050200365]
We propose the first learned multi-frequency image compression and entropy coding approach.
It is based on the recently developed octave convolutions to factorize the latents into high and low frequency (resolution) components.
We show that the proposed generalized octave convolution can improve the performance of other auto-encoder-based computer vision tasks.
arXiv Detail & Related papers (2020-02-24T01:35:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.