PrefGen: Preference Guided Image Generation with Relative Attributes
- URL: http://arxiv.org/abs/2304.00185v1
- Date: Sat, 1 Apr 2023 00:41:51 GMT
- Title: PrefGen: Preference Guided Image Generation with Relative Attributes
- Authors: Alec Helbling, Christopher J. Rozell, Matthew O'Shaughnessy, Kion
Fallah
- Abstract summary: Deep generative models have the capacity to render high fidelity images of content like human faces.
We develop the $textitPrefGen$ system, which allows users to control the relative attributes of generated images.
We demonstrate the success of this approach using a StyleGAN2 generator on the task of human face editing.
- Score: 5.0741409008225755
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Deep generative models have the capacity to render high fidelity images of
content like human faces. Recently, there has been substantial progress in
conditionally generating images with specific quantitative attributes, like the
emotion conveyed by one's face. These methods typically require a user to
explicitly quantify the desired intensity of a visual attribute. A limitation
of this method is that many attributes, like how "angry" a human face looks,
are difficult for a user to precisely quantify. However, a user would be able
to reliably say which of two faces seems "angrier". Following this premise, we
develop the $\textit{PrefGen}$ system, which allows users to control the
relative attributes of generated images by presenting them with simple paired
comparison queries of the form "do you prefer image $a$ or image $b$?" Using
information from a sequence of query responses, we can estimate user
preferences over a set of image attributes and perform preference-guided image
editing and generation. Furthermore, to make preference localization feasible
and efficient, we apply an active query selection strategy. We demonstrate the
success of this approach using a StyleGAN2 generator on the task of human face
editing. Additionally, we demonstrate how our approach can be combined with
CLIP, allowing a user to edit the relative intensity of attributes specified by
text prompts. Code at https://github.com/helblazer811/PrefGen.
Related papers
- ManiCLIP: Multi-Attribute Face Manipulation from Text [104.30600573306991]
We present a novel multi-attribute face manipulation method based on textual descriptions.
Our method generates natural manipulated faces with minimal text-irrelevant attribute editing.
arXiv Detail & Related papers (2022-10-02T07:22:55Z) - FaceController: Controllable Attribute Editing for Face in the Wild [74.56117807309576]
We propose a simple feed-forward network to generate high-fidelity manipulated faces.
By simply employing some existing and easy-obtainable prior information, our method can control, transfer, and edit diverse attributes of faces in the wild.
In our method, we decouple identity, expression, pose, and illumination using 3D priors; separate texture and colors by using region-wise style codes.
arXiv Detail & Related papers (2021-02-23T02:47:28Z) - Attributes Aware Face Generation with Generative Adversarial Networks [133.44359317633686]
We propose a novel attributes aware face image generator method with generative adversarial networks called AFGAN.
Three stacked generators generate $64 times 64$, $128 times 128$ and $256 times 256$ resolution face images respectively.
In addition, an image-attribute matching loss is proposed to enhance the correlation between the generated images and input attributes.
arXiv Detail & Related papers (2020-12-03T09:25:50Z) - S2FGAN: Semantically Aware Interactive Sketch-to-Face Translation [11.724779328025589]
This paper proposes a sketch-to-image generation framework called S2FGAN.
We employ two latent spaces to control the face appearance and adjust the desired attributes of the generated face.
Our method successfully outperforms state-of-the-art methods on attribute manipulation by exploiting greater control of attribute intensity.
arXiv Detail & Related papers (2020-11-30T13:42:39Z) - SMILE: Semantically-guided Multi-attribute Image and Layout Editing [154.69452301122175]
Attribute image manipulation has been a very active topic since the introduction of Generative Adversarial Networks (GANs)
We present a multimodal representation that handles all attributes, be it guided by random noise or images, while only using the underlying domain information of the target domain.
Our method is capable of adding, removing or changing either fine-grained or coarse attributes by using an image as a reference or by exploring the style distribution space.
arXiv Detail & Related papers (2020-10-05T20:15:21Z) - Describe What to Change: A Text-guided Unsupervised Image-to-Image
Translation Approach [84.22327278486846]
We propose a novel unsupervised approach, based on image-to-image translation, that alters the attributes of a given image through a command-like sentence.
Our model disentangles the image content from the visual attributes, and it learns to modify the latter using the textual description.
Experiments show that the proposed model achieves promising performances on two large-scale public datasets.
arXiv Detail & Related papers (2020-08-10T15:40:05Z) - Generating Person Images with Appearance-aware Pose Stylizer [66.44220388377596]
We present a novel end-to-end framework to generate realistic person images based on given person poses and appearances.
The core of our framework is a novel generator called Appearance-aware Pose Stylizer (APS) which generates human images by coupling the target pose with the conditioned person appearance progressively.
arXiv Detail & Related papers (2020-07-17T15:58:05Z) - Conditional Image Generation and Manipulation for User-Specified Content [6.6081578501076494]
We propose a single pipeline for text-to-image generation and manipulation.
In the first part of our pipeline we introduce textStyleGAN, a model that is conditioned on text.
In the second part of our pipeline we make use of the pre-trained weights of textStyleGAN to perform semantic facial image manipulation.
arXiv Detail & Related papers (2020-05-11T08:05:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.