StyleGAN2 Distillation for Feed-forward Image Manipulation
- URL: http://arxiv.org/abs/2003.03581v2
- Date: Thu, 22 Oct 2020 14:07:35 GMT
- Title: StyleGAN2 Distillation for Feed-forward Image Manipulation
- Authors: Yuri Viazovetskyi, Vladimir Ivashkin, Evgeny Kashin
- Abstract summary: StyleGAN2 is a state-of-the-art network in generating realistic images.
We propose a way to distill a particular image manipulation of StyleGAN2 into image-to-image network trained in paired way.
- Score: 5.5080625617632
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: StyleGAN2 is a state-of-the-art network in generating realistic images.
Besides, it was explicitly trained to have disentangled directions in latent
space, which allows efficient image manipulation by varying latent factors.
Editing existing images requires embedding a given image into the latent space
of StyleGAN2. Latent code optimization via backpropagation is commonly used for
qualitative embedding of real world images, although it is prohibitively slow
for many applications. We propose a way to distill a particular image
manipulation of StyleGAN2 into image-to-image network trained in paired way.
The resulting pipeline is an alternative to existing GANs, trained on unpaired
data. We provide results of human faces' transformation: gender swap,
aging/rejuvenation, style transfer and image morphing. We show that the quality
of generation using our method is comparable to StyleGAN2 backpropagation and
current state-of-the-art methods in these particular tasks.
Related papers
- D2Styler: Advancing Arbitrary Style Transfer with Discrete Diffusion Methods [2.468658581089448]
We propose a novel framework called D$2$Styler (Discrete Diffusion Styler)
Our method uses Adaptive Instance Normalization (AdaIN) features as a context guide for the reverse diffusion process.
Experimental results demonstrate that D$2$Styler produces high-quality style-transferred images.
arXiv Detail & Related papers (2024-08-07T05:47:06Z) - Portrait Diffusion: Training-free Face Stylization with
Chain-of-Painting [64.43760427752532]
Face stylization refers to the transformation of a face into a specific portrait style.
Current methods require the use of example-based adaptation approaches to fine-tune pre-trained generative models.
This paper proposes a training-free face stylization framework, named Portrait Diffusion.
arXiv Detail & Related papers (2023-12-03T06:48:35Z) - Improving Diffusion-based Image Translation using Asymmetric Gradient
Guidance [51.188396199083336]
We present an approach that guides the reverse process of diffusion sampling by applying asymmetric gradient guidance.
Our model's adaptability allows it to be implemented with both image-fusion and latent-dif models.
Experiments show that our method outperforms various state-of-the-art models in image translation tasks.
arXiv Detail & Related papers (2023-06-07T12:56:56Z) - Gradient Adjusting Networks for Domain Inversion [82.72289618025084]
StyleGAN2 was demonstrated to be a powerful image generation engine that supports semantic editing.
We present a per-image optimization method that tunes a StyleGAN2 generator such that it achieves a local edit to the generator's weights.
Our experiments show a sizable gap in performance over the current state of the art in this very active domain.
arXiv Detail & Related papers (2023-02-22T14:47:57Z) - StyleT2F: Generating Human Faces from Textual Description Using
StyleGAN2 [0.0]
StyleT2F is a method of controlling the output of StyleGAN2 using text.
Our method proves to capture the required features correctly and shows consistency between the input text and the output images.
arXiv Detail & Related papers (2022-04-17T04:51:30Z) - Saliency Constrained Arbitrary Image Style Transfer using SIFT and DCNN [22.57205921266602]
When common neural style transfer methods are used, the textures and colors in the style image are usually transferred imperfectly to the content image.
This paper proposes a novel saliency constrained method to reduce or avoid such effects.
The experiments show that the saliency maps of source images can help find the correct matching and avoid artifacts.
arXiv Detail & Related papers (2022-01-14T09:00:55Z) - Controllable Person Image Synthesis with Spatially-Adaptive Warped
Normalization [72.65828901909708]
Controllable person image generation aims to produce realistic human images with desirable attributes.
We introduce a novel Spatially-Adaptive Warped Normalization (SAWN), which integrates a learned flow-field to warp modulation parameters.
We propose a novel self-training part replacement strategy to refine the pretrained model for the texture-transfer task.
arXiv Detail & Related papers (2021-05-31T07:07:44Z) - StyleCLIP: Text-Driven Manipulation of StyleGAN Imagery [71.1862388442953]
We develop a text-based interface for StyleGAN image manipulation.
We first introduce an optimization scheme that utilizes a CLIP-based loss to modify an input latent vector in response to a user-provided text prompt.
Next, we describe a latent mapper that infers a text-guided latent manipulation step for a given input image, allowing faster and more stable text-based manipulation.
arXiv Detail & Related papers (2021-03-31T17:51:25Z) - Conditional Image Generation and Manipulation for User-Specified Content [6.6081578501076494]
We propose a single pipeline for text-to-image generation and manipulation.
In the first part of our pipeline we introduce textStyleGAN, a model that is conditioned on text.
In the second part of our pipeline we make use of the pre-trained weights of textStyleGAN to perform semantic facial image manipulation.
arXiv Detail & Related papers (2020-05-11T08:05:00Z) - Exploiting Deep Generative Prior for Versatile Image Restoration and
Manipulation [181.08127307338654]
This work presents an effective way to exploit the image prior captured by a generative adversarial network (GAN) trained on large-scale natural images.
The deep generative prior (DGP) provides compelling results to restore missing semantics, e.g., color, patch, resolution, of various degraded images.
arXiv Detail & Related papers (2020-03-30T17:45:07Z) - P$^2$-GAN: Efficient Style Transfer Using Single Style Image [2.703193151632043]
Style transfer is a useful image synthesis technique that can re-render given image into another artistic style.
Generative Adversarial Network (GAN) is a widely adopted framework toward this task for its better representation ability on local style patterns.
In this paper, a novel Patch Permutation GAN (P$2$-GAN) network that can efficiently learn the stroke style from a single style image is proposed.
arXiv Detail & Related papers (2020-01-21T12:08:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.