GuidedStyle: Attribute Knowledge Guided Style Manipulation for Semantic
Face Editing
- URL: http://arxiv.org/abs/2012.11856v1
- Date: Tue, 22 Dec 2020 06:53:31 GMT
- Title: GuidedStyle: Attribute Knowledge Guided Style Manipulation for Semantic
Face Editing
- Authors: Xianxu Hou, Xiaokang Zhang, Linlin Shen, Zhihui Lai, Jun Wan
- Abstract summary: We propose a novel learning framework, called GuidedStyle, to achieve semantic face editing on StyleGAN.
Our method is able to perform disentangled and controllable edits along various attributes, including smiling, eyeglasses, gender, mustache and hair color.
- Score: 39.57994147985615
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Although significant progress has been made in synthesizing high-quality and
visually realistic face images by unconditional Generative Adversarial Networks
(GANs), there still lacks of control over the generation process in order to
achieve semantic face editing. In addition, it remains very challenging to
maintain other face information untouched while editing the target attributes.
In this paper, we propose a novel learning framework, called GuidedStyle, to
achieve semantic face editing on StyleGAN by guiding the image generation
process with a knowledge network. Furthermore, we allow an attention mechanism
in StyleGAN generator to adaptively select a single layer for style
manipulation. As a result, our method is able to perform disentangled and
controllable edits along various attributes, including smiling, eyeglasses,
gender, mustache and hair color. Both qualitative and quantitative results
demonstrate the superiority of our method over other competing methods for
semantic face editing. Moreover, we show that our model can be also applied to
different types of real and artistic face editing, demonstrating strong
generalization ability.
Related papers
- When StyleGAN Meets Stable Diffusion: a $\mathscr{W}_+$ Adapter for
Personalized Image Generation [60.305112612629465]
Text-to-image diffusion models have excelled in producing diverse, high-quality, and photo-realistic images.
We present a novel use of the extended StyleGAN embedding space $mathcalW_+$ to achieve enhanced identity preservation and disentanglement for diffusion models.
Our method adeptly generates personalized text-to-image outputs that are not only compatible with prompt descriptions but also amenable to common StyleGAN editing directions.
arXiv Detail & Related papers (2023-11-29T09:05:14Z) - DreamIdentity: Improved Editability for Efficient Face-identity
Preserved Image Generation [69.16517915592063]
We propose a novel face-identity encoder to learn an accurate representation of human faces.
We also propose self-augmented editability learning to enhance the editability of models.
Our methods can generate identity-preserved images under different scenes at a much faster speed.
arXiv Detail & Related papers (2023-07-01T11:01:17Z) - ChatFace: Chat-Guided Real Face Editing via Diffusion Latent Space
Manipulation [22.724306705927095]
We propose a novel approach that conduct text-driven image editing in the semantic latent space of diffusion model.
By aligning the temporal feature of the diffusion model with the semantic condition at generative process, we introduce a stable manipulation strategy.
We develop an interactive system named ChatFace, which combines the zero-shot reasoning ability of large language models to perform efficient manipulations.
arXiv Detail & Related papers (2023-05-24T05:28:37Z) - A comprehensive survey on semantic facial attribute editing using
generative adversarial networks [0.688204255655161]
A large number of face generation and manipulation models have been proposed.
Semantic facial attribute editing is the process of varying the values of one or more attributes of a face image.
Based on their architectures, the state-of-the-art models are categorized and studied as encoder-decoder, image-to-image, and photo-guided models.
arXiv Detail & Related papers (2022-05-21T13:09:38Z) - MOST-GAN: 3D Morphable StyleGAN for Disentangled Face Image Manipulation [69.35523133292389]
We propose a framework that a priori models physical attributes of the face explicitly, thus providing disentanglement by design.
Our method, MOST-GAN, integrates the expressive power and photorealism of style-based GANs with the physical disentanglement and flexibility of nonlinear 3D morphable models.
It achieves photorealistic manipulation of portrait images with fully disentangled 3D control over their physical attributes, enabling extreme manipulation of lighting, facial expression, and pose variations up to full profile view.
arXiv Detail & Related papers (2021-11-01T15:53:36Z) - FacialGAN: Style Transfer and Attribute Manipulation on Synthetic Faces [9.664892091493586]
FacialGAN is a novel framework enabling simultaneous rich style transfers and interactive facial attributes manipulation.
We show our model's capacity in producing visually compelling results in style transfer, attribute manipulation, diversity and face verification.
arXiv Detail & Related papers (2021-10-18T15:53:38Z) - FaceController: Controllable Attribute Editing for Face in the Wild [74.56117807309576]
We propose a simple feed-forward network to generate high-fidelity manipulated faces.
By simply employing some existing and easy-obtainable prior information, our method can control, transfer, and edit diverse attributes of faces in the wild.
In our method, we decouple identity, expression, pose, and illumination using 3D priors; separate texture and colors by using region-wise style codes.
arXiv Detail & Related papers (2021-02-23T02:47:28Z) - Enjoy Your Editing: Controllable GANs for Image Editing via Latent Space
Navigation [136.53288628437355]
Controllable semantic image editing enables a user to change entire image attributes with few clicks.
Current approaches often suffer from attribute edits that are entangled, global image identity changes, and diminished photo-realism.
We propose quantitative evaluation strategies for measuring controllable editing performance, unlike prior work which primarily focuses on qualitative evaluation.
arXiv Detail & Related papers (2021-02-01T21:38:36Z) - S2FGAN: Semantically Aware Interactive Sketch-to-Face Translation [11.724779328025589]
This paper proposes a sketch-to-image generation framework called S2FGAN.
We employ two latent spaces to control the face appearance and adjust the desired attributes of the generated face.
Our method successfully outperforms state-of-the-art methods on attribute manipulation by exploiting greater control of attribute intensity.
arXiv Detail & Related papers (2020-11-30T13:42:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.