One-shot domain adaptation for semantic face editing of real world
images using StyleALAE
- URL: http://arxiv.org/abs/2108.13876v1
- Date: Tue, 31 Aug 2021 14:32:18 GMT
- Title: One-shot domain adaptation for semantic face editing of real world
images using StyleALAE
- Authors: Ravi Kiran Reddy, Kumar Shubham, Gopalakrishnan Venkatesh, Sriram
Gandikota, Sarthak Khoche, Dinesh Babu Jayagopi, Gopalakrishnan
Srinivasaraghavan
- Abstract summary: styleALAE is a latent-space based autoencoder that can generate photo-realistic images of high quality.
Our work ensures that the identity of the reconstructed image is the same as the given input image.
We further generate semantic modifications over the reconstructed image by using the latent space of the pre-trained styleALAE model.
- Score: 7.541747299649292
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Semantic face editing of real world facial images is an important application
of generative models. Recently, multiple works have explored possible
techniques to generate such modifications using the latent structure of
pre-trained GAN models. However, such approaches often require training an
encoder network and that is typically a time-consuming and resource intensive
process. A possible alternative to such a GAN-based architecture can be
styleALAE, a latent-space based autoencoder that can generate photo-realistic
images of high quality. Unfortunately, the reconstructed image in styleALAE
does not preserve the identity of the input facial image. This limits the
application of styleALAE for semantic face editing of images with known
identities. In our work, we use a recent advancement in one-shot domain
adaptation to address this problem. Our work ensures that the identity of the
reconstructed image is the same as the given input image. We further generate
semantic modifications over the reconstructed image by using the latent space
of the pre-trained styleALAE model. Results show that our approach can generate
semantic modifications on any real world facial image while preserving the
identity.
Related papers
- Semantic Image Synthesis via Class-Adaptive Cross-Attention [7.147779225315707]
Cross-attention layers are used in place of SPADE for learning shape-style correlations and so conditioning the image generation process.
Our model inherits the versatility of SPADE, at the same time obtaining state-of-the-art generation quality, as well as improved global and local style transfer.
arXiv Detail & Related papers (2023-08-30T14:49:34Z) - DreamIdentity: Improved Editability for Efficient Face-identity
Preserved Image Generation [69.16517915592063]
We propose a novel face-identity encoder to learn an accurate representation of human faces.
We also propose self-augmented editability learning to enhance the editability of models.
Our methods can generate identity-preserved images under different scenes at a much faster speed.
arXiv Detail & Related papers (2023-07-01T11:01:17Z) - Semantics-Guided Object Removal for Facial Images: with Broad
Applicability and Robust Style Preservation [29.162655333387452]
Object removal and image inpainting in facial images is a task in which objects that occlude a facial image are specifically targeted, removed, and replaced by a properly reconstructed facial image.
Two different approaches utilizing U-net and modulated generator respectively have been widely endorsed for this task for their unique advantages but notwithstanding each method's innate disadvantages.
Here, we propose Semantics-Guided Inpainting Network (SGIN) which itself is a modification of the modulated generator, aiming to take advantage of its advanced generative capability and preserve the high-fidelity details of the original image.
arXiv Detail & Related papers (2022-09-29T00:09:12Z) - Controllable Person Image Synthesis with Spatially-Adaptive Warped
Normalization [72.65828901909708]
Controllable person image generation aims to produce realistic human images with desirable attributes.
We introduce a novel Spatially-Adaptive Warped Normalization (SAWN), which integrates a learned flow-field to warp modulation parameters.
We propose a novel self-training part replacement strategy to refine the pretrained model for the texture-transfer task.
arXiv Detail & Related papers (2021-05-31T07:07:44Z) - Inverting Generative Adversarial Renderer for Face Reconstruction [58.45125455811038]
In this work, we introduce a novel Generative Adversa Renderer (GAR)
GAR learns to model the complicated real-world image, instead of relying on the graphics rules, it is capable of producing realistic images.
Our method achieves state-of-the-art performances on multiple face reconstruction.
arXiv Detail & Related papers (2021-05-06T04:16:06Z) - Ensembling with Deep Generative Views [72.70801582346344]
generative models can synthesize "views" of artificial images that mimic real-world variations, such as changes in color or pose.
Here, we investigate whether such views can be applied to real images to benefit downstream analysis tasks such as image classification.
We use StyleGAN2 as the source of generative augmentations and investigate this setup on classification tasks involving facial attributes, cat faces, and cars.
arXiv Detail & Related papers (2021-04-29T17:58:35Z) - Enjoy Your Editing: Controllable GANs for Image Editing via Latent Space
Navigation [136.53288628437355]
Controllable semantic image editing enables a user to change entire image attributes with few clicks.
Current approaches often suffer from attribute edits that are entangled, global image identity changes, and diminished photo-realism.
We propose quantitative evaluation strategies for measuring controllable editing performance, unlike prior work which primarily focuses on qualitative evaluation.
arXiv Detail & Related papers (2021-02-01T21:38:36Z) - S2FGAN: Semantically Aware Interactive Sketch-to-Face Translation [11.724779328025589]
This paper proposes a sketch-to-image generation framework called S2FGAN.
We employ two latent spaces to control the face appearance and adjust the desired attributes of the generated face.
Our method successfully outperforms state-of-the-art methods on attribute manipulation by exploiting greater control of attribute intensity.
arXiv Detail & Related papers (2020-11-30T13:42:39Z) - Generating Person Images with Appearance-aware Pose Stylizer [66.44220388377596]
We present a novel end-to-end framework to generate realistic person images based on given person poses and appearances.
The core of our framework is a novel generator called Appearance-aware Pose Stylizer (APS) which generates human images by coupling the target pose with the conditioned person appearance progressively.
arXiv Detail & Related papers (2020-07-17T15:58:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.