PA-GAN: Progressive Attention Generative Adversarial Network for Facial
Attribute Editing
- URL: http://arxiv.org/abs/2007.05892v1
- Date: Sun, 12 Jul 2020 03:04:12 GMT
- Title: PA-GAN: Progressive Attention Generative Adversarial Network for Facial
Attribute Editing
- Authors: Zhenliang He, Meina Kan, Jichao Zhang, Shiguang Shan
- Abstract summary: We propose a progressive attention GAN (PA-GAN) for facial attribute editing.
Our approach achieves correct attribute editing with irrelevant details much better preserved compared with the state-of-the-arts.
- Score: 67.94255549416548
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Facial attribute editing aims to manipulate attributes on the human face,
e.g., adding a mustache or changing the hair color. Existing approaches suffer
from a serious compromise between correct attribute generation and preservation
of the other information such as identity and background, because they edit the
attributes in the imprecise area. To resolve this dilemma, we propose a
progressive attention GAN (PA-GAN) for facial attribute editing. In our
approach, the editing is progressively conducted from high to low feature level
while being constrained inside a proper attribute area by an attention mask at
each level. This manner prevents undesired modifications to the irrelevant
regions from the beginning, and then the network can focus more on correctly
generating the attributes within a proper boundary at each level. As a result,
our approach achieves correct attribute editing with irrelevant details much
better preserved compared with the state-of-the-arts. Codes are released at
https://github.com/LynnHo/PA-GAN-Tensorflow.
Related papers
- EVA: Zero-shot Accurate Attributes and Multi-Object Video Editing [62.15822650722473]
Current video editing methods fail to edit the foreground and background simultaneously while preserving the original layout.
We introduce EVA, a textbfzero-shot and textbfmulti-attribute video editing framework tailored for human-centric videos with complex motions.
EVA can be easily generalized to multi-object editing scenarios and achieves accurate identity mapping.
arXiv Detail & Related papers (2024-03-24T12:04:06Z) - Identity-preserving Editing of Multiple Facial Attributes by Learning
Global Edit Directions and Local Adjustments [4.082799056366928]
ID-Style is a new architecture capable of addressing the problem of identity loss during attribute manipulation.
We introduce two losses during training to enforce the LGD to find semi-sparse semantic directions, which along with the IAIP, preserve the identity of the input instance.
arXiv Detail & Related papers (2023-09-25T16:28:39Z) - Face Attribute Editing with Disentangled Latent Vectors [0.0]
We propose an image-to-image translation framework for facial attribute editing.
Inspired by the latent space factorization works of fixed pretrained GANs, we design the attribute editing by latent space factorization.
To project images to semantically organized latent spaces, we set an encoder-decoder architecture with attention-based skip connections.
arXiv Detail & Related papers (2023-01-11T18:32:13Z) - ManiCLIP: Multi-Attribute Face Manipulation from Text [104.30600573306991]
We present a novel multi-attribute face manipulation method based on textual descriptions.
Our method generates natural manipulated faces with minimal text-irrelevant attribute editing.
arXiv Detail & Related papers (2022-10-02T07:22:55Z) - PASS: Protected Attribute Suppression System for Mitigating Bias in Face
Recognition [55.858374644761525]
Face recognition networks encode information about sensitive attributes while being trained for identity classification.
Existing bias mitigation approaches require end-to-end training and are unable to achieve high verification accuracy.
We present a descriptors-based adversarial de-biasing approach called Protected Attribute Suppression System ( PASS)'
Pass can be trained on top of descriptors obtained from any previously trained high-performing network to classify identities and simultaneously reduce encoding of sensitive attributes.
arXiv Detail & Related papers (2021-08-09T00:39:22Z) - FaceController: Controllable Attribute Editing for Face in the Wild [74.56117807309576]
We propose a simple feed-forward network to generate high-fidelity manipulated faces.
By simply employing some existing and easy-obtainable prior information, our method can control, transfer, and edit diverse attributes of faces in the wild.
In our method, we decouple identity, expression, pose, and illumination using 3D priors; separate texture and colors by using region-wise style codes.
arXiv Detail & Related papers (2021-02-23T02:47:28Z) - CAFE-GAN: Arbitrary Face Attribute Editing with Complementary Attention
Feature [31.425326840578098]
We propose a novel GAN model which is designed to edit only the parts of a face pertinent to the target attributes.
CAFE identifies the facial regions to be transformed by considering both target attributes as well as complementary attributes.
arXiv Detail & Related papers (2020-11-24T05:21:03Z) - MagGAN: High-Resolution Face Attribute Editing with Mask-Guided
Generative Adversarial Network [145.4591079418917]
MagGAN learns to only edit the facial parts that are relevant to the desired attribute changes.
A novel mask-guided conditioning strategy is introduced to incorporate the influence region of each attribute change into the generator.
A multi-level patch-wise discriminator structure is proposed to scale our model for high-resolution ($1024 times 1024$) face editing.
arXiv Detail & Related papers (2020-10-03T20:56:16Z) - MU-GAN: Facial Attribute Editing based on Multi-attention Mechanism [12.762892831902349]
We propose a Multi-attention U-Net-based Generative Adversarial Network (MU-GAN)
First, we replace a classic convolutional encoder-decoder with a symmetric U-Net-like structure in a generator.
Second, a self-attention mechanism is incorporated into convolutional layers for modeling long-range and multi-level dependencies.
arXiv Detail & Related papers (2020-09-09T09:25:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.