Contextual Residual Aggregation for Ultra High-Resolution Image
Inpainting
- URL: http://arxiv.org/abs/2005.09704v1
- Date: Tue, 19 May 2020 18:55:32 GMT
- Title: Contextual Residual Aggregation for Ultra High-Resolution Image
Inpainting
- Authors: Zili Yi, Qiang Tang, Shekoofeh Azizi, Daesik Jang, Zhan Xu
- Abstract summary: We propose a Contextual Residual Aggregation (CRA) mechanism that can produce high-frequency residuals for missing contents.
CRA mechanism produces high-frequency residuals for missing contents by weighted aggregating residuals from contextual patches.
We train the proposed model on small images with resolutions 512x512 and perform inference on high-resolution images, achieving compelling inpainting quality.
- Score: 12.839962012888199
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recently data-driven image inpainting methods have made inspiring progress,
impacting fundamental image editing tasks such as object removal and damaged
image repairing. These methods are more effective than classic approaches,
however, due to memory limitations they can only handle low-resolution inputs,
typically smaller than 1K. Meanwhile, the resolution of photos captured with
mobile devices increases up to 8K. Naive up-sampling of the low-resolution
inpainted result can merely yield a large yet blurry result. Whereas, adding a
high-frequency residual image onto the large blurry image can generate a sharp
result, rich in details and textures. Motivated by this, we propose a
Contextual Residual Aggregation (CRA) mechanism that can produce high-frequency
residuals for missing contents by weighted aggregating residuals from
contextual patches, thus only requiring a low-resolution prediction from the
network. Since convolutional layers of the neural network only need to operate
on low-resolution inputs and outputs, the cost of memory and computing power is
thus well suppressed. Moreover, the need for high-resolution training datasets
is alleviated. In our experiments, we train the proposed model on small images
with resolutions 512x512 and perform inference on high-resolution images,
achieving compelling inpainting quality. Our model can inpaint images as large
as 8K with considerable hole sizes, which is intractable with previous
learning-based approaches. We further elaborate on the light-weight design of
the network architecture, achieving real-time performance on 2K images on a GTX
1080 Ti GPU. Codes are available at: Atlas200dk/sample-imageinpainting-HiFill.
Related papers
- CoordFill: Efficient High-Resolution Image Inpainting via Parameterized
Coordinate Querying [52.91778151771145]
In this paper, we try to break the limitations for the first time thanks to the recent development of continuous implicit representation.
Experiments show that the proposed method achieves real-time performance on the 2048$times$2048 images using a single GTX 2080 Ti GPU.
arXiv Detail & Related papers (2023-03-15T11:13:51Z) - LR-Net: A Block-based Convolutional Neural Network for Low-Resolution
Image Classification [0.0]
We develop a novel image classification architecture, composed of blocks that are designed to learn both low level and global features from noisy and low-resolution images.
Our design of the blocks was heavily influenced by Residual Connections and Inception modules in order to increase performance and reduce parameter sizes.
We have performed in-depth tests that demonstrate the presented architecture is faster and more accurate than existing cutting-edge convolutional neural networks.
arXiv Detail & Related papers (2022-07-19T20:01:11Z) - Learning Enriched Features for Fast Image Restoration and Enhancement [166.17296369600774]
This paper presents a holistic goal of maintaining spatially-precise high-resolution representations through the entire network.
We learn an enriched set of features that combines contextual information from multiple scales, while simultaneously preserving the high-resolution spatial details.
Our approach achieves state-of-the-art results for a variety of image processing tasks, including defocus deblurring, image denoising, super-resolution, and image enhancement.
arXiv Detail & Related papers (2022-04-19T17:59:45Z) - Restormer: Efficient Transformer for High-Resolution Image Restoration [118.9617735769827]
convolutional neural networks (CNNs) perform well at learning generalizable image priors from large-scale data.
Transformers have shown significant performance gains on natural language and high-level vision tasks.
Our model, named Restoration Transformer (Restormer), achieves state-of-the-art results on several image restoration tasks.
arXiv Detail & Related papers (2021-11-18T18:59:10Z) - SDWNet: A Straight Dilated Network with Wavelet Transformation for Image
Deblurring [23.86692375792203]
Image deblurring is a computer vision problem that aims to recover a sharp image from a blurred image.
Our model uses dilated convolution to enable the obtainment of the large receptive field with high spatial resolution.
We propose a novel module using the wavelet transform, which effectively helps the network to recover clear high-frequency texture details.
arXiv Detail & Related papers (2021-10-12T07:58:10Z) - Generating Superpixels for High-resolution Images with Decoupled Patch
Calibration [82.21559299694555]
Patch Networks (PCNet) is designed to efficiently and accurately implement high-resolution superpixel segmentation.
DPC takes a local patch from the high-resolution images and dynamically generates a binary mask to impose the network to focus on region boundaries.
In particular, DPC takes a local patch from the high-resolution images and dynamically generates a binary mask to impose the network to focus on region boundaries.
arXiv Detail & Related papers (2021-08-19T10:33:05Z) - Exploiting Raw Images for Real-Scene Super-Resolution [105.18021110372133]
We study the problem of real-scene single image super-resolution to bridge the gap between synthetic data and real captured images.
We propose a method to generate more realistic training data by mimicking the imaging process of digital cameras.
We also develop a two-branch convolutional neural network to exploit the radiance information originally-recorded in raw images.
arXiv Detail & Related papers (2021-02-02T16:10:15Z) - Invertible Image Rescaling [118.2653765756915]
We develop an Invertible Rescaling Net (IRN) to produce visually-pleasing low-resolution images.
We capture the distribution of the lost information using a latent variable following a specified distribution in the downscaling process.
arXiv Detail & Related papers (2020-05-12T09:55:53Z) - Deep Attentive Generative Adversarial Network for Photo-Realistic Image
De-Quantization [25.805568996596783]
De-quantization can improve the visual quality of low bit-depth image to display on high bit-depth screen.
This paper proposes DAGAN algorithm to perform super-resolution on image intensity resolution.
DenseResAtt module consists of dense residual blocks armed with self-attention mechanism.
arXiv Detail & Related papers (2020-04-07T06:45:01Z) - Perceptual Image Super-Resolution with Progressive Adversarial Network [17.289101902846358]
Single Image Super-Resolution (SISR) aims to improve resolution of small-size low-quality image from a single one.
In this paper, we argue that the curse of dimensionality is the underlying reason of limiting the performance of state-of-the-art algorithms.
We propose Progressive Adversarial Network (PAN) that is capable of coping with this difficulty for domain-specific image super-resolution.
arXiv Detail & Related papers (2020-03-08T10:19:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.