PoNA: Pose-guided Non-local Attention for Human Pose Transfer
- URL: http://arxiv.org/abs/2012.07049v1
- Date: Sun, 13 Dec 2020 12:38:29 GMT
- Title: PoNA: Pose-guided Non-local Attention for Human Pose Transfer
- Authors: Kun Li, Jinsong Zhang, Yebin Liu, Yu-Kun Lai, Qionghai Dai
- Abstract summary: We propose a new human pose transfer method using a generative adversarial network (GAN) with simplified cascaded blocks.
Our model generates sharper and more realistic images with rich details, while having fewer parameters and faster speed.
- Score: 105.14398322129024
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Human pose transfer, which aims at transferring the appearance of a given
person to a target pose, is very challenging and important in many
applications. Previous work ignores the guidance of pose features or only uses
local attention mechanism, leading to implausible and blurry results. We
propose a new human pose transfer method using a generative adversarial network
(GAN) with simplified cascaded blocks. In each block, we propose a pose-guided
non-local attention (PoNA) mechanism with a long-range dependency scheme to
select more important regions of image features to transfer. We also design
pre-posed image-guided pose feature update and post-posed pose-guided image
feature update to better utilize the pose and image features. Our network is
simple, stable, and easy to train. Quantitative and qualitative results on
Market-1501 and DeepFashion datasets show the efficacy and efficiency of our
model. Compared with state-of-the-art methods, our model generates sharper and
more realistic images with rich details, while having fewer parameters and
faster speed. Furthermore, our generated images can help to alleviate data
insufficiency for person re-identification.
Related papers
- DiffUHaul: A Training-Free Method for Object Dragging in Images [78.93531472479202]
We propose a training-free method, dubbed DiffUHaul, for the object dragging task.
We first apply attention masking in each denoising step to make the generation more disentangled across different objects.
In the early denoising steps, we interpolate the attention features between source and target images to smoothly fuse new layouts with the original appearance.
arXiv Detail & Related papers (2024-06-03T17:59:53Z) - Pose Guided Human Image Synthesis with Partially Decoupled GAN [25.800174118151638]
Pose Guided Human Image Synthesis (PGHIS) is a challenging task of transforming a human image from the reference pose to a target pose.
We propose a method by decoupling the human body into several parts to guide the synthesis of a realistic image of the person.
In addition, we design a multi-head attention-based module for PGHIS.
arXiv Detail & Related papers (2022-10-07T15:31:37Z) - Pose-driven Attention-guided Image Generation for Person
Re-Identification [39.605062525247135]
We propose an end-to-end pose-driven generative adversarial network to generate multiple poses of a person.
A semantic-consistency loss is proposed to preserve the semantic information of the person during pose transfer.
We show that by incorporating the proposed approach in a person re-identification framework, realistic pose transferred images and state-of-the-art re-identification results can be achieved.
arXiv Detail & Related papers (2021-04-28T14:02:24Z) - Learning Semantic Person Image Generation by Region-Adaptive
Normalization [81.52223606284443]
We propose a new two-stage framework to handle the pose and appearance translation.
In the first stage, we predict the target semantic parsing maps to eliminate the difficulties of pose transfer.
In the second stage, we suggest a new person image generation method by incorporating the region-adaptive normalization.
arXiv Detail & Related papers (2021-04-14T06:51:37Z) - Progressive and Aligned Pose Attention Transfer for Person Image
Generation [59.87492938953545]
This paper proposes a new generative adversarial network for pose transfer, i.e., transferring the pose of a given person to a target pose.
We use two types of blocks, namely Pose-Attentional Transfer Block (PATB) and Aligned Pose-Attentional Transfer Bloc (APATB)
We verify the efficacy of the model on the Market-1501 and DeepFashion datasets, using quantitative and qualitative measures.
arXiv Detail & Related papers (2021-03-22T07:24:57Z) - MUST-GAN: Multi-level Statistics Transfer for Self-driven Person Image
Generation [13.06676286691587]
Pose-guided person image generation usually involves using paired source-target images to supervise the training.
We propose a novel multi-level statistics transfer model, which disentangles and transfers multi-level appearance features from person images.
Our approach allows for flexible manipulation of person appearance and pose properties to perform pose transfer and clothes style transfer tasks.
arXiv Detail & Related papers (2020-11-18T04:38:48Z) - Person image generation with semantic attention network for person
re-identification [9.30413920076019]
We propose a novel person pose-guided image generation method, which is called the semantic attention network.
The network consists of several semantic attention blocks, where each block attends to preserve and update the pose code and the clothing textures.
Compared with other methods, our network can characterize better body shape and keep clothing attributes, simultaneously.
arXiv Detail & Related papers (2020-08-18T12:18:51Z) - Adversarially-Trained Deep Nets Transfer Better: Illustration on Image
Classification [53.735029033681435]
Transfer learning is a powerful methodology for adapting pre-trained deep neural networks on image recognition tasks to new domains.
In this work, we demonstrate that adversarially-trained models transfer better than non-adversarially-trained models.
arXiv Detail & Related papers (2020-07-11T22:48:42Z) - Neural Pose Transfer by Spatially Adaptive Instance Normalization [73.04483812364127]
We propose the first neural pose transfer model that solves the pose transfer via the latest technique for image style transfer.
Our model does not require any correspondences between the source and target meshes.
Experiments show that the proposed model can effectively transfer deformation from source to target meshes, and has good generalization ability to deal with unseen identities or poses of meshes.
arXiv Detail & Related papers (2020-03-16T14:33:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.