VCNet: A Robust Approach to Blind Image Inpainting
- URL: http://arxiv.org/abs/2003.06816v1
- Date: Sun, 15 Mar 2020 12:47:57 GMT
- Title: VCNet: A Robust Approach to Blind Image Inpainting
- Authors: Yi Wang, Ying-Cong Chen, Xin Tao, Jiaya Jia
- Abstract summary: Blind inpainting is a task to automatically complete visual contents without specifying masks for missing areas in an image.
In this paper, we define a new blind inpainting setting, making training a blind inpainting neural system robust against unknown missing region patterns.
Our method is effective and robust in blind image inpainting. And our VCN allows for a wide spectrum of applications.
- Score: 70.68227719731243
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Blind inpainting is a task to automatically complete visual contents without
specifying masks for missing areas in an image. Previous works assume missing
region patterns are known, limiting its application scope. In this paper, we
relax the assumption by defining a new blind inpainting setting, making
training a blind inpainting neural system robust against various unknown
missing region patterns. Specifically, we propose a two-stage visual
consistency network (VCN), meant to estimate where to fill (via masks) and
generate what to fill. In this procedure, the unavoidable potential mask
prediction errors lead to severe artifacts in the subsequent repairing. To
address it, our VCN predicts semantically inconsistent regions first, making
mask prediction more tractable. Then it repairs these estimated missing regions
using a new spatial normalization, enabling VCN to be robust to the mask
prediction errors. In this way, semantically convincing and visually compelling
content is thus generated. Extensive experiments are conducted, showing our
method is effective and robust in blind image inpainting. And our VCN allows
for a wide spectrum of applications.
Related papers
- Towards Improved Input Masking for Convolutional Neural Networks [66.99060157800403]
We propose a new masking method for CNNs we call layer masking.
We show that our method is able to eliminate or minimize the influence of the mask shape or color on the output of the model.
We also demonstrate how the shape of the mask may leak information about the class, thus affecting estimates of model reliance on class-relevant features.
arXiv Detail & Related papers (2022-11-26T19:31:49Z) - Learning Prior Feature and Attention Enhanced Image Inpainting [63.21231753407192]
This paper incorporates the pre-training based Masked AutoEncoder (MAE) into the inpainting model.
We propose to use attention priors from MAE to make the inpainting model learn more long-distance dependencies between masked and unmasked regions.
arXiv Detail & Related papers (2022-08-03T04:32:53Z) - NeRF-In: Free-Form NeRF Inpainting with RGB-D Priors [14.911856302172996]
We introduce the first framework that enables users to remove unwanted objects or undesired regions in a 3D scene represented by a pre-trained NeRF.
We show it obtained visual plausible and structurally consistent results across multiple views using shorter time and less user manual efforts.
arXiv Detail & Related papers (2022-06-10T06:54:22Z) - RePaint: Inpainting using Denoising Diffusion Probabilistic Models [161.74792336127345]
Free-form inpainting is the task of adding new content to an image in the regions specified by an arbitrary binary mask.
We propose RePaint: A Denoising Probabilistic Model (DDPM) based inpainting approach that is applicable to even extreme masks.
We validate our method for both faces and general-purpose image inpainting using standard and extreme masks.
arXiv Detail & Related papers (2022-01-24T18:40:15Z) - FT-TDR: Frequency-guided Transformer and Top-Down Refinement Network for
Blind Face Inpainting [77.78305705925376]
Blind face inpainting refers to the task of reconstructing visual contents without explicitly indicating the corrupted regions in a face image.
We propose a novel two-stage blind face inpainting method named Frequency-guided Transformer and Top-Down Refinement Network (FT-TDR) to tackle these challenges.
arXiv Detail & Related papers (2021-08-10T03:12:01Z) - Image Inpainting by End-to-End Cascaded Refinement with Mask Awareness [66.55719330810547]
Inpainting arbitrary missing regions is challenging because learning valid features for various masked regions is nontrivial.
We propose a novel mask-aware inpainting solution that learns multi-scale features for missing regions in the encoding phase.
Our framework is validated both quantitatively and qualitatively via extensive experiments on three public datasets.
arXiv Detail & Related papers (2021-04-28T13:17:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.