GLaMa: Joint Spatial and Frequency Loss for General Image Inpainting
- URL: http://arxiv.org/abs/2205.07162v1
- Date: Sun, 15 May 2022 02:18:59 GMT
- Title: GLaMa: Joint Spatial and Frequency Loss for General Image Inpainting
- Authors: Zeyu Lu and Junjun Jiang and Junqin Huang and Gang Wu and Xianming Liu
- Abstract summary: The purpose of image inpainting is to recover scratches and damaged areas using context information from remaining parts.
We propose a simple yet general method to solve this problem based on the LaMa image inpainting framework, dubbed GLaMa.
Our proposed GLaMa can better capture different types of missing information by using more types of masks.
- Score: 44.04779984090629
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The purpose of image inpainting is to recover scratches and damaged areas
using context information from remaining parts. In recent years, thanks to the
resurgence of convolutional neural networks (CNNs), image inpainting task has
made great breakthroughs. However, most of the work consider insufficient types
of mask, and their performance will drop dramatically when encountering unseen
masks. To combat these challenges, we propose a simple yet general method to
solve this problem based on the LaMa image inpainting framework, dubbed GLaMa.
Our proposed GLaMa can better capture different types of missing information by
using more types of masks. By incorporating more degraded images in the
training phase, we can expect to enhance the robustness of the model with
respect to various masks. In order to yield more reasonable results, we further
introduce a frequency-based loss in addition to the traditional spatial
reconstruction loss and adversarial loss. In particular, we introduce an
effective reconstruction loss both in the spatial and frequency domain to
reduce the chessboard effect and ripples in the reconstructed image. Extensive
experiments demonstrate that our method can boost the performance over the
original LaMa method for each type of mask on FFHQ, ImageNet, Places2 and
WikiArt dataset. The proposed GLaMa was ranked first in terms of PSNR, LPIPS
and SSIM in the NTIRE 2022 Image Inpainting Challenge Track 1 Unsupervised.
Related papers
- Restore Anything with Masks: Leveraging Mask Image Modeling for Blind All-in-One Image Restoration [35.3663995646582]
All-in-one image restoration aims to handle multiple degradation types using one model.
This paper proposes a simple pipeline for all-in-one blind image restoration to Restore Anything with Masks (RAM)
arXiv Detail & Related papers (2024-09-28T16:33:43Z) - Face Mask Removal with Region-attentive Face Inpainting [0.7433327915285965]
We propose a generative face inpainting method to recover/reconstruct the masked part of a face.
Our proposed method includes a Multi-scale Channel-Spatial Attention Module (M-CSAM) to mitigate the spatial information loss.
We synthesize our own Masked-Faces dataset from the CelebA dataset by incorporating five different types of face masks.
arXiv Detail & Related papers (2024-09-10T20:10:11Z) - DPPMask: Masked Image Modeling with Determinantal Point Processes [49.65141962357528]
Masked Image Modeling (MIM) has achieved impressive representative performance with the aim of reconstructing randomly masked images.
We show that uniformly random masking widely used in previous works unavoidably loses some key objects and changes original semantic information.
To address this issue, we augment MIM with a new masking strategy namely the DPPMask.
Our method is simple yet effective and requires no extra learnable parameters when implemented within various frameworks.
arXiv Detail & Related papers (2023-03-13T13:40:39Z) - Masked Face Inpainting Through Residual Attention UNet [0.7868449549351486]
This paper proposes a blind mask face inpainting method using residual attention UNet.
A residual block feeds info to the next layer and directly into the layers about two hops away to solve the vanishing gradient problem.
Experiments on the publicly available CelebA dataset show the feasibility and robustness of our proposed model.
arXiv Detail & Related papers (2022-09-19T08:49:53Z) - Learning Prior Feature and Attention Enhanced Image Inpainting [63.21231753407192]
This paper incorporates the pre-training based Masked AutoEncoder (MAE) into the inpainting model.
We propose to use attention priors from MAE to make the inpainting model learn more long-distance dependencies between masked and unmasked regions.
arXiv Detail & Related papers (2022-08-03T04:32:53Z) - Shape-Aware Masking for Inpainting in Medical Imaging [49.61617087640379]
Inpainting has been proposed as a successful deep learning technique for unsupervised medical image model discovery.
We introduce a method for generating shape-aware masks for inpainting, which aims at learning the statistical shape prior.
We propose an unsupervised guided masking approach based on an off-the-shelf inpainting model and a superpixel over-segmentation algorithm.
arXiv Detail & Related papers (2022-07-12T18:35:17Z) - RePaint: Inpainting using Denoising Diffusion Probabilistic Models [161.74792336127345]
Free-form inpainting is the task of adding new content to an image in the regions specified by an arbitrary binary mask.
We propose RePaint: A Denoising Probabilistic Model (DDPM) based inpainting approach that is applicable to even extreme masks.
We validate our method for both faces and general-purpose image inpainting using standard and extreme masks.
arXiv Detail & Related papers (2022-01-24T18:40:15Z) - Resolution-robust Large Mask Inpainting with Fourier Convolutions [10.152370311844445]
Inpainting systems often struggle with large missing areas, complex geometric structures, and high-resolution images.
We find that one of the main reasons for that is the lack of an effective receptive field in both the inpainting network and the loss function.
We propose a new method called large mask inpainting (LaMa) to alleviate this issue.
arXiv Detail & Related papers (2021-09-15T08:54:29Z) - Image Inpainting by End-to-End Cascaded Refinement with Mask Awareness [66.55719330810547]
Inpainting arbitrary missing regions is challenging because learning valid features for various masked regions is nontrivial.
We propose a novel mask-aware inpainting solution that learns multi-scale features for missing regions in the encoding phase.
Our framework is validated both quantitatively and qualitatively via extensive experiments on three public datasets.
arXiv Detail & Related papers (2021-04-28T13:17:47Z) - DeepGIN: Deep Generative Inpainting Network for Extreme Image Inpainting [45.39552853543588]
We propose a deep generative inpainting network, named DeepGIN, to handle various types of masked images.
Our model is capable of completing masked images in the wild.
arXiv Detail & Related papers (2020-08-17T09:30:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.