DrawingInStyles: Portrait Image Generation and Editing with Spatially
Conditioned StyleGAN
- URL: http://arxiv.org/abs/2203.02762v1
- Date: Sat, 5 Mar 2022 14:54:07 GMT
- Title: DrawingInStyles: Portrait Image Generation and Editing with Spatially
Conditioned StyleGAN
- Authors: Wanchao Su, Hui Ye, Shu-Yu Chen, Lin Gao, Hongbo Fu
- Abstract summary: We introduce SC-StyleGAN, which injects spatial constraints to the original StyleGAN generation process.
Based on SC-StyleGAN, we present DrawingInStyles, a novel drawing interface for non-professional users to easily produce high-quality, photo-realistic face images.
- Score: 30.465955123686335
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The research topic of sketch-to-portrait generation has witnessed a boost of
progress with deep learning techniques. The recently proposed StyleGAN
architectures achieve state-of-the-art generation ability but the original
StyleGAN is not friendly for sketch-based creation due to its unconditional
generation nature. To address this issue, we propose a direct conditioning
strategy to better preserve the spatial information under the StyleGAN
framework. Specifically, we introduce Spatially Conditioned StyleGAN
(SC-StyleGAN for short), which explicitly injects spatial constraints to the
original StyleGAN generation process. We explore two input modalities, sketches
and semantic maps, which together allow users to express desired generation
results more precisely and easily. Based on SC-StyleGAN, we present
DrawingInStyles, a novel drawing interface for non-professional users to easily
produce high-quality, photo-realistic face images with precise control, either
from scratch or editing existing ones. Qualitative and quantitative evaluations
show the superior generation ability of our method to existing and alternative
solutions. The usability and expressiveness of our system are confirmed by a
user study.
Related papers
- PS-StyleGAN: Illustrative Portrait Sketching using Attention-Based Style Adaptation [0.0]
Portrait sketching involves capturing identity specific attributes of a real face with abstract lines and shades.
This paper introduces textbfPortrait Sketching StyleGAN (PS-StyleGAN), a style transfer approach tailored for portrait sketch synthesis.
We leverage the semantic $W+$ latent space of StyleGAN to generate portrait sketches, allowing us to make meaningful edits, like pose and expression alterations, without compromising identity.
arXiv Detail & Related papers (2024-08-31T04:22:45Z) - Deformable One-shot Face Stylization via DINO Semantic Guidance [12.771707124161665]
This paper addresses the issue of one-shot face stylization, focusing on the simultaneous consideration of appearance and structure.
We explore deformation-aware face stylization that diverges from traditional single-image style reference, opting for a real-style image pair instead.
arXiv Detail & Related papers (2024-03-01T11:30:55Z) - HiCAST: Highly Customized Arbitrary Style Transfer with Adapter Enhanced
Diffusion Models [84.12784265734238]
The goal of Arbitrary Style Transfer (AST) is injecting the artistic features of a style reference into a given image/video.
We propose HiCAST, which is capable of explicitly customizing the stylization results according to various source of semantic clues.
A novel learning objective is leveraged for video diffusion model training, which significantly improve cross-frame temporal consistency.
arXiv Detail & Related papers (2024-01-11T12:26:23Z) - Face Cartoonisation For Various Poses Using StyleGAN [0.7673339435080445]
This paper presents an innovative approach to achieve face cartoonisation while preserving the original identity and accommodating various poses.
We achieve this by introducing an encoder that captures both pose and identity information from images and generates a corresponding embedding within the StyleGAN latent space.
We show by extensive experimentation how our encoder adapts the StyleGAN output to better preserve identity when the objective is cartoonisation.
arXiv Detail & Related papers (2023-09-26T13:10:25Z) - StyleAdapter: A Unified Stylized Image Generation Model [97.24936247688824]
StyleAdapter is a unified stylized image generation model capable of producing a variety of stylized images.
It can be integrated with existing controllable synthesis methods, such as T2I-adapter and ControlNet.
arXiv Detail & Related papers (2023-09-04T19:16:46Z) - StyleSwap: Style-Based Generator Empowers Robust Face Swapping [90.05775519962303]
We introduce a concise and effective framework named StyleSwap.
Our core idea is to leverage a style-based generator to empower high-fidelity and robust face swapping.
We identify that with only minimal modifications, a StyleGAN2 architecture can successfully handle the desired information from both source and target.
arXiv Detail & Related papers (2022-09-27T16:35:16Z) - Adversarial Style Augmentation for Domain Generalized Urban-Scene
Segmentation [120.96012935286913]
We propose a novel adversarial style augmentation approach, which can generate hard stylized images during training.
Experiments on two synthetic-to-real semantic segmentation benchmarks demonstrate that AdvStyle can significantly improve the model performance on unseen real domains.
arXiv Detail & Related papers (2022-07-11T14:01:25Z) - Pastiche Master: Exemplar-Based High-Resolution Portrait Style Transfer [103.54337984566877]
Recent studies on StyleGAN show high performance on artistic portrait generation by transfer learning with limited data.
We introduce a novel DualStyleGAN with flexible control of dual styles of the original face domain and the extended artistic portrait domain.
Experiments demonstrate the superiority of DualStyleGAN over state-of-the-art methods in high-quality portrait style transfer and flexible style control.
arXiv Detail & Related papers (2022-03-24T17:57:11Z) - State-of-the-Art in the Architecture, Methods and Applications of
StyleGAN [41.359306557243706]
State-of-the-art report covers the StyleGAN architecture, and the ways it has been employed since its conception.
StyleGAN's learned latent space is surprisingly well-behaved and remarkably disentangled.
The control offered by StyleGAN is inherently limited to the generator's learned distribution.
arXiv Detail & Related papers (2022-02-28T18:42:04Z) - BlendGAN: Implicitly GAN Blending for Arbitrary Stylized Face Generation [9.370501805054344]
We propose BlendGAN for arbitrary stylized face generation.
We first train a self-supervised style encoder on the generic artistic dataset to extract the representations of arbitrary styles.
In addition, a weighted blending module (WBM) is proposed to blend face and style representations implicitly and control the arbitrary stylization effect.
arXiv Detail & Related papers (2021-10-22T12:00:27Z) - Drafting and Revision: Laplacian Pyramid Network for Fast High-Quality
Artistic Style Transfer [115.13853805292679]
Artistic style transfer aims at migrating the style from an example image to a content image.
Inspired by the common painting process of drawing a draft and revising the details, we introduce a novel feed-forward method named Laplacian Pyramid Network (LapStyle)
Our method can synthesize high quality stylized images in real time, where holistic style patterns are properly transferred.
arXiv Detail & Related papers (2021-04-12T11:53:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.