Face Generation and Editing with StyleGAN: A Survey
- URL: http://arxiv.org/abs/2212.09102v3
- Date: Wed, 27 Sep 2023 14:34:46 GMT
- Title: Face Generation and Editing with StyleGAN: A Survey
- Authors: Andrew Melnik, Maksim Miasayedzenkau, Dzianis Makarovets, Dzianis
Pirshtuk, Eren Akbulut, Dennis Holzmann, Tarek Renusch, Gustav Reichert,
Helge Ritter
- Abstract summary: The survey covers the evolution of StyleGAN, from PGGAN to StyleGAN3, and explores relevant topics such as suitable metrics for training, different latent representations, GAN inversion to latent spaces of StyleGAN, face image editing, cross-domain face stylization, face restoration, and even Deepfake applications.
- Score: 0.12362527696478205
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Our goal with this survey is to provide an overview of the state of the art
deep learning methods for face generation and editing using StyleGAN. The
survey covers the evolution of StyleGAN, from PGGAN to StyleGAN3, and explores
relevant topics such as suitable metrics for training, different latent
representations, GAN inversion to latent spaces of StyleGAN, face image
editing, cross-domain face stylization, face restoration, and even Deepfake
applications. We aim to provide an entry point into the field for readers that
have basic knowledge about the field of deep learning and are looking for an
accessible introduction and overview.
Related papers
- Revealing Directions for Text-guided 3D Face Editing [52.85632020601518]
3D face editing is a significant task in multimedia, aimed at the manipulation of 3D face models across various control signals.
We present Face Clan, a text-general approach for generating and manipulating 3D faces based on arbitrary attribute descriptions.
Our method offers a precisely controllable manipulation method, allowing users to intuitively customize regions of interest with the text description.
arXiv Detail & Related papers (2024-10-07T12:04:39Z) - VitaGlyph: Vitalizing Artistic Typography with Flexible Dual-branch Diffusion Models [53.59400446543756]
We introduce a dual-branch and training-free method, namely VitaGlyph, to enable flexible artistic typography.
VitaGlyph treats input character as a scene composed of Subject and Surrounding, followed by rendering them under varying degrees of geometry transformation.
Experimental results demonstrate that VitaGlyph not only achieves better artistry and readability, but also manages to depict multiple customize concepts.
arXiv Detail & Related papers (2024-10-02T16:48:47Z) - Fashion Style Editing with Generative Human Prior [9.854813629782681]
In this work, we aim to manipulate the fashion style of human imagery using text descriptions.
Specifically, we leverage a generative human prior and achieve fashion style editing by navigating its learned latent space.
Our framework successfully projects abstract fashion concepts onto human images and introduces exciting new applications to the field.
arXiv Detail & Related papers (2024-04-02T14:22:04Z) - Face Cartoonisation For Various Poses Using StyleGAN [0.7673339435080445]
This paper presents an innovative approach to achieve face cartoonisation while preserving the original identity and accommodating various poses.
We achieve this by introducing an encoder that captures both pose and identity information from images and generates a corresponding embedding within the StyleGAN latent space.
We show by extensive experimentation how our encoder adapts the StyleGAN output to better preserve identity when the objective is cartoonisation.
arXiv Detail & Related papers (2023-09-26T13:10:25Z) - Few-Shot Font Generation by Learning Fine-Grained Local Styles [90.39288370855115]
Few-shot font generation (FFG) aims to generate a new font with a few examples.
We propose a new font generation approach by learning 1) the fine-grained local styles from references, and 2) the spatial correspondence between the content and reference glyphs.
arXiv Detail & Related papers (2022-05-20T05:07:05Z) - DrawingInStyles: Portrait Image Generation and Editing with Spatially
Conditioned StyleGAN [30.465955123686335]
We introduce SC-StyleGAN, which injects spatial constraints to the original StyleGAN generation process.
Based on SC-StyleGAN, we present DrawingInStyles, a novel drawing interface for non-professional users to easily produce high-quality, photo-realistic face images.
arXiv Detail & Related papers (2022-03-05T14:54:07Z) - Styleverse: Towards Identity Stylization across Heterogeneous Domains [70.13327076710269]
We propose a new challenging task namely IDentity Stylization (IDS) across heterogeneous domains.
We use an effective heterogeneous-network-based framework $Styleverse$ that uses a single domain-aware generator.
$Styleverse achieves higher-fidelity identity stylization compared with other state-of-the-art methods.
arXiv Detail & Related papers (2022-03-02T04:23:01Z) - State-of-the-Art in the Architecture, Methods and Applications of
StyleGAN [41.359306557243706]
State-of-the-art report covers the StyleGAN architecture, and the ways it has been employed since its conception.
StyleGAN's learned latent space is surprisingly well-behaved and remarkably disentangled.
The control offered by StyleGAN is inherently limited to the generator's learned distribution.
arXiv Detail & Related papers (2022-02-28T18:42:04Z) - FEAT: Face Editing with Attention [70.89233432407305]
We build on the StyleGAN generator and present a method that explicitly encourages face manipulation to focus on the intended regions.
During the generation of the edited image, the attention map serves as a mask that guides a blending between the original features and the modified ones.
arXiv Detail & Related papers (2022-02-06T06:07:34Z) - A Brief Survey on Deep Learning Based Data Hiding, Steganography and
Watermarking [98.1953404873897]
We conduct a brief yet comprehensive review of existing literature and outline three meta-architectures.
Based on this, we summarize specific strategies for various applications of deep hiding, including steganography, light field messaging and watermarking.
arXiv Detail & Related papers (2021-03-02T10:01:03Z) - GuidedStyle: Attribute Knowledge Guided Style Manipulation for Semantic
Face Editing [39.57994147985615]
We propose a novel learning framework, called GuidedStyle, to achieve semantic face editing on StyleGAN.
Our method is able to perform disentangled and controllable edits along various attributes, including smiling, eyeglasses, gender, mustache and hair color.
arXiv Detail & Related papers (2020-12-22T06:53:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.