StyleFlow: Attribute-conditioned Exploration of StyleGAN-Generated
Images using Conditional Continuous Normalizing Flows
- URL: http://arxiv.org/abs/2008.02401v2
- Date: Sun, 20 Sep 2020 15:39:46 GMT
- Title: StyleFlow: Attribute-conditioned Exploration of StyleGAN-Generated
Images using Conditional Continuous Normalizing Flows
- Authors: Rameen Abdal, Peihao Zhu, Niloy Mitra, Peter Wonka
- Abstract summary: StyleFlow is an instance of conditional continuous normalizing flows in the GAN latent space conditioned by attribute features.
We evaluate our method using the face and the car latent space of StyleGAN, and demonstrate fine-grained disentangled edits along various attributes on both real photographs and StyleGAN generated images.
- Score: 40.69516201141587
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: High-quality, diverse, and photorealistic images can now be generated by
unconditional GANs (e.g., StyleGAN). However, limited options exist to control
the generation process using (semantic) attributes, while still preserving the
quality of the output. Further, due to the entangled nature of the GAN latent
space, performing edits along one attribute can easily result in unwanted
changes along other attributes. In this paper, in the context of conditional
exploration of entangled latent spaces, we investigate the two sub-problems of
attribute-conditioned sampling and attribute-controlled editing. We present
StyleFlow as a simple, effective, and robust solution to both the sub-problems
by formulating conditional exploration as an instance of conditional continuous
normalizing flows in the GAN latent space conditioned by attribute features. We
evaluate our method using the face and the car latent space of StyleGAN, and
demonstrate fine-grained disentangled edits along various attributes on both
real photographs and StyleGAN generated images. For example, for faces, we vary
camera pose, illumination variation, expression, facial hair, gender, and age.
Finally, via extensive qualitative and quantitative comparisons, we demonstrate
the superiority of StyleFlow to other concurrent works.
Related papers
- ZePo: Zero-Shot Portrait Stylization with Faster Sampling [61.14140480095604]
This paper presents an inversion-free portrait stylization framework based on diffusion models that accomplishes content and style feature fusion in merely four sampling steps.
We propose a feature merging strategy to amalgamate redundant features in Consistency Features, thereby reducing the computational load of attention control.
arXiv Detail & Related papers (2024-08-10T08:53:41Z) - Hierarchical Semantic Regularization of Latent Spaces in StyleGANs [53.98170188547775]
We propose a Hierarchical Semantic Regularizer (HSR) which aligns the hierarchical representations learnt by the generator to corresponding powerful features learnt by pretrained networks on large amounts of data.
HSR is shown to not only improve generator representations but also the linearity and smoothness of the latent style spaces, leading to the generation of more natural-looking style-edited images.
arXiv Detail & Related papers (2022-08-07T16:23:33Z) - Everything is There in Latent Space: Attribute Editing and Attribute
Style Manipulation by StyleGAN Latent Space Exploration [39.18239951479647]
We present Few-shot Latent-based Attribute Manipulation and Editing (FLAME)
FLAME is a framework to perform highly controlled image editing by latent space manipulation.
We generate diverse attribute styles in disentangled manner.
arXiv Detail & Related papers (2022-07-20T12:40:32Z) - DyStyle: Dynamic Neural Network for Multi-Attribute-Conditioned Style
Editing [12.80013698957431]
A Dynamic Style Manipulation Network (DyStyle) is proposed to perform attribute-conditioned style editing.
A novel easy-to-hard training procedure is introduced for efficient and stable training of the DyStyle network.
Our approach demonstrates fine-grained disentangled edits along multiple numeric and binary attributes.
arXiv Detail & Related papers (2021-09-22T13:50:51Z) - Disentangled Face Attribute Editing via Instance-Aware Latent Space
Search [30.17338705964925]
A rich set of semantic directions exist in the latent space of Generative Adversarial Networks (GANs)
Existing methods may suffer poor attribute variation disentanglement, leading to unwanted change of other attributes when altering the desired one.
We propose a novel framework (IALS) that performs Instance-Aware Latent-Space Search to find semantic directions for disentangled attribute editing.
arXiv Detail & Related papers (2021-05-26T16:19:08Z) - Enjoy Your Editing: Controllable GANs for Image Editing via Latent Space
Navigation [136.53288628437355]
Controllable semantic image editing enables a user to change entire image attributes with few clicks.
Current approaches often suffer from attribute edits that are entangled, global image identity changes, and diminished photo-realism.
We propose quantitative evaluation strategies for measuring controllable editing performance, unlike prior work which primarily focuses on qualitative evaluation.
arXiv Detail & Related papers (2021-02-01T21:38:36Z) - Style Intervention: How to Achieve Spatial Disentanglement with
Style-based Generators? [100.60938767993088]
We propose a lightweight optimization-based algorithm which could adapt to arbitrary input images and render natural translation effects under flexible objectives.
We verify the performance of the proposed framework in facial attribute editing on high-resolution images, where both photo-realism and consistency are required.
arXiv Detail & Related papers (2020-11-19T07:37:31Z) - PIE: Portrait Image Embedding for Semantic Control [82.69061225574774]
We present the first approach for embedding real portrait images in the latent space of StyleGAN.
We use StyleRig, a pretrained neural network that maps the control space of a 3D morphable face model to the latent space of the GAN.
An identity energy preservation term allows spatially coherent edits while maintaining facial integrity.
arXiv Detail & Related papers (2020-09-20T17:53:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.