Large Hole Image Inpainting With Compress-Decompression Network
- URL: http://arxiv.org/abs/2002.00199v1
- Date: Sat, 1 Feb 2020 12:39:13 GMT
- Title: Large Hole Image Inpainting With Compress-Decompression Network
- Authors: Zhenghang Wu, Yidong Cui
- Abstract summary: Existing methods propose convolutional neural networks to repair corrupted images.
We study the existing approaches and propose a new network, the compression-decompression network.
We construct the compression network with the residual network and propose a similar texture selection algorithm to extend the image.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Image inpainting technology can patch images with missing pixels. Existing
methods propose convolutional neural networks to repair corrupted images. The
networks focus on the valid pixels around the missing pixels, use the
encoder-decoder structure to extract valuable information, and use the
information to fix the vacancy. However, if the missing part is too large to
provide useful information, the result will exist blur, color mixing, and
object confusion. In order to patch the large hole image, we study the existing
approaches and propose a new network, the compression-decompression network.
The compression network takes responsibility for inpainting and generating a
down-sample image. The decompression network takes responsibility for extending
the down-sample image into the original resolution. We construct the
compression network with the residual network and propose a similar texture
selection algorithm to extend the image that is better than using the
super-resolution network. We evaluate our model over Places2 and CelebA data
set and use the similarity ratio as the metric. The result shows that our model
has better performance when the inpainting task has many conflicts.
Related papers
- Perceptual Image Compression with Cooperative Cross-Modal Side
Information [53.356714177243745]
We propose a novel deep image compression method with text-guided side information to achieve a better rate-perception-distortion tradeoff.
Specifically, we employ the CLIP text encoder and an effective Semantic-Spatial Aware block to fuse the text and image features.
arXiv Detail & Related papers (2023-11-23T08:31:11Z) - You Can Mask More For Extremely Low-Bitrate Image Compression [80.7692466922499]
Learned image compression (LIC) methods have experienced significant progress during recent years.
LIC methods fail to explicitly explore the image structure and texture components crucial for image compression.
We present DA-Mask that samples visible patches based on the structure and texture of original images.
We propose a simple yet effective masked compression model (MCM), the first framework that unifies LIC and LIC end-to-end for extremely low-bitrate compression.
arXiv Detail & Related papers (2023-06-27T15:36:22Z) - Analysis of the Effect of Low-Overhead Lossy Image Compression on the
Performance of Visual Crowd Counting for Smart City Applications [78.55896581882595]
Lossy image compression techniques can reduce the quality of the images, leading to accuracy degradation.
In this paper, we analyze the effect of applying low-overhead lossy image compression methods on the accuracy of visual crowd counting.
arXiv Detail & Related papers (2022-07-20T19:20:03Z) - Estimating the Resize Parameter in End-to-end Learned Image Compression [50.20567320015102]
We describe a search-free resizing framework that can further improve the rate-distortion tradeoff of recent learned image compression models.
Our results show that our new resizing parameter estimation framework can provide Bjontegaard-Delta rate (BD-rate) improvement of about 10% against leading perceptual quality engines.
arXiv Detail & Related papers (2022-04-26T01:35:02Z) - The Devil Is in the Details: Window-based Attention for Image
Compression [58.1577742463617]
Most existing learned image compression models are based on Convolutional Neural Networks (CNNs)
In this paper, we study the effects of multiple kinds of attention mechanisms for local features learning, then introduce a more straightforward yet effective window-based local attention block.
The proposed window-based attention is very flexible which could work as a plug-and-play component to enhance CNN and Transformer models.
arXiv Detail & Related papers (2022-03-16T07:55:49Z) - Image Inpainting Using AutoEncoder and Guided Selection of Predicted
Pixels [9.527576103168984]
In this paper, we propose a network for image inpainting. This network, similar to U-Net, extracts various features from images, leading to better results.
We improved the final results by replacing the damaged pixels with the recovered pixels of the output images.
arXiv Detail & Related papers (2021-12-17T00:10:34Z) - SDWNet: A Straight Dilated Network with Wavelet Transformation for Image
Deblurring [23.86692375792203]
Image deblurring is a computer vision problem that aims to recover a sharp image from a blurred image.
Our model uses dilated convolution to enable the obtainment of the large receptive field with high spatial resolution.
We propose a novel module using the wavelet transform, which effectively helps the network to recover clear high-frequency texture details.
arXiv Detail & Related papers (2021-10-12T07:58:10Z) - Image Compression with Recurrent Neural Network and Generalized Divisive
Normalization [3.0204520109309843]
Deep learning has gained huge attention from the research community and produced promising image reconstruction results.
Recent methods focused on developing deeper and more complex networks, which significantly increased network complexity.
In this paper, two effective novel blocks are developed: analysis and block synthesis that employs the convolution layer and Generalized Divisive Normalization (GDN) in the variable-rate encoder and decoder side.
arXiv Detail & Related papers (2021-09-05T05:31:55Z) - Variable-Rate Deep Image Compression through Spatially-Adaptive Feature
Transform [58.60004238261117]
We propose a versatile deep image compression network based on Spatial Feature Transform (SFT arXiv:1804.02815)
Our model covers a wide range of compression rates using a single model, which is controlled by arbitrary pixel-wise quality maps.
The proposed framework allows us to perform task-aware image compressions for various tasks.
arXiv Detail & Related papers (2021-08-21T17:30:06Z) - Image Complexity Guided Network Compression for Biomedical Image
Segmentation [5.926887379656135]
We propose an image complexity-guided network compression technique for biomedical image segmentation.
We map the dataset complexity to the target network accuracy degradation caused by compression.
The mapping is used to determine the convolutional layer-wise multiplicative factor for generating a compressed network.
arXiv Detail & Related papers (2021-07-06T22:28:10Z) - R-MNet: A Perceptual Adversarial Network for Image Inpainting [5.471225956329675]
We propose a Wasserstein GAN combined with a new reverse mask operator, namely Reverse Masking Network (R-MNet), a perceptual adversarial network for image inpainting.
We show that our method is able to generalize to high-resolution inpainting task, and further show more realistic outputs that are plausible to the human visual system.
arXiv Detail & Related papers (2020-08-11T10:58:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.