AniGAN: Style-Guided Generative Adversarial Networks for Unsupervised
Anime Face Generation
- URL: http://arxiv.org/abs/2102.12593v1
- Date: Wed, 24 Feb 2021 22:47:38 GMT
- Title: AniGAN: Style-Guided Generative Adversarial Networks for Unsupervised
Anime Face Generation
- Authors: Bing Li, Yuanlue Zhu, Yitong Wang, Chia-Wen Lin, Bernard Ghanem,
Linlin Shen
- Abstract summary: We propose a novel framework to translate a portrait photo-face into an anime appearance.
Our aim is to synthesize anime-faces which are style-consistent with a given reference anime-face.
Existing methods often fail to transfer the styles of reference anime-faces, or introduce noticeable artifacts/distortions in the local shapes of their generated faces.
- Score: 84.52819242283852
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we propose a novel framework to translate a portrait
photo-face into an anime appearance. Our aim is to synthesize anime-faces which
are style-consistent with a given reference anime-face. However, unlike typical
translation tasks, such anime-face translation is challenging due to complex
variations of appearances among anime-faces. Existing methods often fail to
transfer the styles of reference anime-faces, or introduce noticeable
artifacts/distortions in the local shapes of their generated faces. We propose
Ani- GAN, a novel GAN-based translator that synthesizes highquality
anime-faces. Specifically, a new generator architecture is proposed to
simultaneously transfer color/texture styles and transform local facial shapes
into anime-like counterparts based on the style of a reference anime-face,
while preserving the global structure of the source photoface. We propose a
double-branch discriminator to learn both domain-specific distributions and
domain-shared distributions, helping generate visually pleasing anime-faces and
effectively mitigate artifacts. Extensive experiments qualitatively and
quantitatively demonstrate the superiority of our method over state-of-the-art
methods.
Related papers
- FlashFace: Human Image Personalization with High-fidelity Identity Preservation [59.76645602354481]
FlashFace allows users to easily personalize their own photos by providing one or a few reference face images and a text prompt.
Our approach is distinguishable from existing human photo customization methods by higher-fidelity identity preservation and better instruction following.
arXiv Detail & Related papers (2024-03-25T17:59:57Z) - When StyleGAN Meets Stable Diffusion: a $\mathscr{W}_+$ Adapter for
Personalized Image Generation [60.305112612629465]
Text-to-image diffusion models have excelled in producing diverse, high-quality, and photo-realistic images.
We present a novel use of the extended StyleGAN embedding space $mathcalW_+$ to achieve enhanced identity preservation and disentanglement for diffusion models.
Our method adeptly generates personalized text-to-image outputs that are not only compatible with prompt descriptions but also amenable to common StyleGAN editing directions.
arXiv Detail & Related papers (2023-11-29T09:05:14Z) - Language-Guided Face Animation by Recurrent StyleGAN-based Generator [87.56260982475564]
We study a novel task, language-guided face animation, that aims to animate a static face image with the help of languages.
We propose a recurrent motion generator to extract a series of semantic and motion information from the language and feed it along with visual information to a pre-trained StyleGAN to generate high-quality frames.
arXiv Detail & Related papers (2022-08-11T02:57:30Z) - 3D Cartoon Face Generation with Controllable Expressions from a Single
GAN Image [142.047662926209]
We generate 3D cartoon face shapes from single 2D GAN generated human faces.
We manipulate latent codes to generate images with different poses and lighting, such that we can reconstruct the 3D cartoon face shapes.
arXiv Detail & Related papers (2022-07-29T01:06:21Z) - Cross-Domain Style Mixing for Face Cartoonization [6.174413879403037]
We propose a novel method called Cross-domain Style mixing, which combines two latent codes from two different domains.
Our method effectively stylizes faces into multiple cartoon characters at various face abstraction levels using only a single generator.
arXiv Detail & Related papers (2022-05-25T02:39:10Z) - Anime-to-Real Clothing: Cosplay Costume Generation via Image-to-Image
Translation [2.4660652494309936]
This paper presents an automatic costume image generation method based on image-to-image translation.
We present a novel architecture for generative adversarial networks (GANs) to facilitate high-quality cosplay image generation.
Experiments demonstrated that, with two types of evaluation metrics, the proposed GAN achieves better performance than existing methods.
arXiv Detail & Related papers (2020-08-26T10:34:46Z) - MangaGAN: Unpaired Photo-to-Manga Translation Based on The Methodology
of Manga Drawing [27.99490750445691]
We propose MangaGAN, the first method based on Generative Adversarial Network (GAN) for unpaired photo-to-manga translation.
Inspired by how experienced manga artists draw manga, MangaGAN generates the geometric features of manga face by a designed GAN model.
To produce high-quality manga faces, we propose a structural smoothing loss to smooth stroke-lines and avoid noisy pixels, and a similarity preserving module.
arXiv Detail & Related papers (2020-04-22T15:23:42Z) - MW-GAN: Multi-Warping GAN for Caricature Generation with Multi-Style
Geometric Exaggeration [53.98437317161086]
Given an input face photo, the goal of caricature generation is to produce stylized, exaggerated caricatures that share the same identity as the photo.
We propose a novel framework called Multi-Warping GAN (MW-GAN), including a style network and a geometric network.
Experiments show that caricatures generated by MW-GAN have better quality than existing methods.
arXiv Detail & Related papers (2020-01-07T03:08:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.