High-Resolution GAN Inversion for Degraded Images in Large Diverse
Datasets
- URL: http://arxiv.org/abs/2302.03406v1
- Date: Tue, 7 Feb 2023 11:24:11 GMT
- Title: High-Resolution GAN Inversion for Degraded Images in Large Diverse
Datasets
- Authors: Yanbo Wang, Chuming Lin, Donghao Luo, Ying Tai, Zhizhong Zhang, Yuan
Xie
- Abstract summary: In this paper, we present a novel GAN inversion framework that utilizes the powerful generative ability of StyleGAN-XL.
To ease the inversion challenge with StyleGAN-XL, Clustering & Regularize Inversion (CRI) is proposed.
We validate our CRI scheme on multiple restoration tasks (i.e., inpainting, colorization, and super-resolution) of complex natural images, and show preferable quantitative and qualitative results.
- Score: 39.21692649763314
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The last decades are marked by massive and diverse image data, which shows
increasingly high resolution and quality. However, some images we obtained may
be corrupted, affecting the perception and the application of downstream tasks.
A generic method for generating a high-quality image from the degraded one is
in demand. In this paper, we present a novel GAN inversion framework that
utilizes the powerful generative ability of StyleGAN-XL for this problem. To
ease the inversion challenge with StyleGAN-XL, Clustering \& Regularize
Inversion (CRI) is proposed. Specifically, the latent space is firstly divided
into finer-grained sub-spaces by clustering. Instead of initializing the
inversion with the average latent vector, we approximate a centroid latent
vector from the clusters, which generates an image close to the input image.
Then, an offset with a regularization term is introduced to keep the inverted
latent vector within a certain range. We validate our CRI scheme on multiple
restoration tasks (i.e., inpainting, colorization, and super-resolution) of
complex natural images, and show preferable quantitative and qualitative
results. We further demonstrate our technique is robust in terms of data and
different GAN models. To our best knowledge, we are the first to adopt
StyleGAN-XL for generating high-quality natural images from diverse degraded
inputs. Code is available at https://github.com/Booooooooooo/CRI.
Related papers
- Solving General Noisy Inverse Problem via Posterior Sampling: A Policy Gradient Viewpoint [21.22750301965104]
We leverage a pretrained diffusion generative model to solve a wide range of image inverse tasks without task specific model fine-tuning.
To precisely estimate the guidance score function of the input image, we propose Diffusion Policy Gradient (DPG)
Experiments show that our method is robust to both Gaussian and Poisson noise degradation on multiple linear and non-linear inverse tasks.
arXiv Detail & Related papers (2024-03-15T16:38:47Z) - High-Fidelity Image Inpainting with GAN Inversion [23.49170140410603]
In this paper, we propose a novel GAN inversion model for image inpainting, dubbed InvertFill.
Within the encoder, the pre-modulation network leverages multi-scale structures to encode more discriminative semantics into style vectors.
To reconstruct faithful and photorealistic images, a simple yet effective Soft-update Mean Latent module is designed to capture more diverse in-domain patterns that synthesize high-fidelity textures for large corruptions.
arXiv Detail & Related papers (2022-08-25T03:39:24Z) - InfinityGAN: Towards Infinite-Resolution Image Synthesis [92.40782797030977]
We present InfinityGAN, a method to generate arbitrary-resolution images.
We show how it trains and infers patch-by-patch seamlessly with low computational resources.
arXiv Detail & Related papers (2021-04-08T17:59:30Z) - Adversarial Generation of Continuous Images [31.92891885615843]
In this paper, we propose two novel architectural techniques for building INR-based image decoders.
We use them to build a state-of-the-art continuous image GAN.
Our proposed INR-GAN architecture improves the performance of continuous image generators by several times.
arXiv Detail & Related papers (2020-11-24T11:06:40Z) - A Contrastive Learning Approach for Training Variational Autoencoder
Priors [137.62674958536712]
Variational autoencoders (VAEs) are one of the powerful likelihood-based generative models with applications in many domains.
One explanation for VAEs' poor generative quality is the prior hole problem: the prior distribution fails to match the aggregate approximate posterior.
We propose an energy-based prior defined by the product of a base prior distribution and a reweighting factor, designed to bring the base closer to the aggregate posterior.
arXiv Detail & Related papers (2020-10-06T17:59:02Z) - Exploiting Deep Generative Prior for Versatile Image Restoration and
Manipulation [181.08127307338654]
This work presents an effective way to exploit the image prior captured by a generative adversarial network (GAN) trained on large-scale natural images.
The deep generative prior (DGP) provides compelling results to restore missing semantics, e.g., color, patch, resolution, of various degraded images.
arXiv Detail & Related papers (2020-03-30T17:45:07Z) - PULSE: Self-Supervised Photo Upsampling via Latent Space Exploration of
Generative Models [77.32079593577821]
PULSE (Photo Upsampling via Latent Space Exploration) generates high-resolution, realistic images at resolutions previously unseen in the literature.
Our method outperforms state-of-the-art methods in perceptual quality at higher resolutions and scale factors than previously possible.
arXiv Detail & Related papers (2020-03-08T16:44:31Z) - Image Fine-grained Inpainting [89.17316318927621]
We present a one-stage model that utilizes dense combinations of dilated convolutions to obtain larger and more effective receptive fields.
To better train this efficient generator, except for frequently-used VGG feature matching loss, we design a novel self-guided regression loss.
We also employ a discriminator with local and global branches to ensure local-global contents consistency.
arXiv Detail & Related papers (2020-02-07T03:45:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.