LI-Net: Large-Pose Identity-Preserving Face Reenactment Network
- URL: http://arxiv.org/abs/2104.02850v1
- Date: Wed, 7 Apr 2021 01:41:21 GMT
- Title: LI-Net: Large-Pose Identity-Preserving Face Reenactment Network
- Authors: Jin Liu, Peng Chen, Tao Liang, Zhaoxing Li, Cai Yu, Shuqiao Zou, Jiao
Dai, Jizhong Han
- Abstract summary: We propose a large-pose identity-preserving face reenactment network, LI-Net.
Specifically, the Landmark Transformer is adopted to adjust driving landmark images.
The Face Rotation Module and the Expression Enhancing Generator decouple the transformed landmark image into pose and expression features, and reenact those attributes separately to generate identity-preserving faces.
- Score: 14.472453602392182
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Face reenactment is a challenging task, as it is difficult to maintain
accurate expression, pose and identity simultaneously. Most existing methods
directly apply driving facial landmarks to reenact source faces and ignore the
intrinsic gap between two identities, resulting in the identity mismatch issue.
Besides, they neglect the entanglement of expression and pose features when
encoding driving faces, leading to inaccurate expressions and visual artifacts
on large-pose reenacted faces. To address these problems, we propose a
Large-pose Identity-preserving face reenactment network, LI-Net. Specifically,
the Landmark Transformer is adopted to adjust driving landmark images, which
aims to narrow the identity gap between driving and source landmark images.
Then the Face Rotation Module and the Expression Enhancing Generator decouple
the transformed landmark image into pose and expression features, and reenact
those attributes separately to generate identity-preserving faces with accurate
expressions and poses. Both qualitative and quantitative experimental results
demonstrate the superiority of our method.
Related papers
- AniFaceDiff: High-Fidelity Face Reenactment via Facial Parametric Conditioned Diffusion Models [33.39336530229545]
Face reenactment refers to the process of transferring the pose and facial expressions from a reference (driving) video onto a static facial (source) image.
Previous research in this domain has made significant progress by training controllable deep generative models to generate faces.
This paper proposes a new method based on Stable Diffusion, called AniFaceDiff, incorporating a new conditioning module for high-fidelity face reenactment.
arXiv Detail & Related papers (2024-06-19T07:08:48Z) - HyperReenact: One-Shot Reenactment via Jointly Learning to Refine and
Retarget Faces [47.27033282706179]
We present our method for neural face reenactment, called HyperReenact, that aims to generate realistic talking head images of a source identity.
Our method operates under the one-shot setting (i.e., using a single source frame) and allows for cross-subject reenactment, without requiring subject-specific fine-tuning.
We compare our method both quantitatively and qualitatively against several state-of-the-art techniques on the standard benchmarks of VoxCeleb1 and VoxCeleb2.
arXiv Detail & Related papers (2023-07-20T11:59:42Z) - Face Transformer: Towards High Fidelity and Accurate Face Swapping [54.737909435708936]
Face swapping aims to generate swapped images that fuse the identity of source faces and the attributes of target faces.
This paper presents Face Transformer, a novel face swapping network that can accurately preserve source identities and target attributes simultaneously.
arXiv Detail & Related papers (2023-04-05T15:51:44Z) - Semantic-aware One-shot Face Re-enactment with Dense Correspondence
Estimation [100.60938767993088]
One-shot face re-enactment is a challenging task due to the identity mismatch between source and driving faces.
This paper proposes to use 3D Morphable Model (3DMM) for explicit facial semantic decomposition and identity disentanglement.
arXiv Detail & Related papers (2022-11-23T03:02:34Z) - StyleMask: Disentangling the Style Space of StyleGAN2 for Neural Face
Reenactment [47.27033282706179]
We propose a framework that learns to disentangle the identity characteristics of the face from its pose.
We show that the proposed method produces higher quality results even on extreme pose variations.
arXiv Detail & Related papers (2022-09-27T13:22:35Z) - Disentangling Identity and Pose for Facial Expression Recognition [54.50747989860957]
We propose an identity and pose disentangled facial expression recognition (IPD-FER) model to learn more discriminative feature representation.
For identity encoder, a well pre-trained face recognition model is utilized and fixed during training, which alleviates the restriction on specific expression training data.
By comparing the difference between synthesized neutral and expressional images of the same individual, the expression component is further disentangled from identity and pose.
arXiv Detail & Related papers (2022-08-17T06:48:13Z) - Graph-based Generative Face Anonymisation with Pose Preservation [49.18049578591058]
AnonyGAN is a GAN-based solution for face anonymisation.
It replaces the visual information corresponding to a source identity with a condition identity provided as any single image.
arXiv Detail & Related papers (2021-12-10T12:58:17Z) - FACEGAN: Facial Attribute Controllable rEenactment GAN [24.547319786399743]
Face reenactment is a popular animation method where the person's identity is taken from the source image and the facial motion from the driving image.
Recent works have demonstrated high quality results by combining the facial landmark based motion representations with the generative adversarial networks.
We propose a novel Facial Attribute Controllable rEenactment GAN (FACEGAN), which transfers the facial motion from the driving face via the Action Unit (AU) representation.
arXiv Detail & Related papers (2020-11-09T14:04:15Z) - FaR-GAN for One-Shot Face Reenactment [20.894596219099164]
We present a one-shot face reenactment model, FaR-GAN, that takes only one face image of any given source identity and a target expression as input.
The proposed method makes no assumptions about the source identity, facial expression, head pose, or even image background.
arXiv Detail & Related papers (2020-05-13T16:15:37Z) - One-Shot Identity-Preserving Portrait Reenactment [16.889479797252783]
We present a deep learning-based framework for portrait reenactment from a single picture of a target (one-shot) and a video of a driving subject.
We aim to address identity preservation in cross-subject portrait reenactment from a single picture.
arXiv Detail & Related papers (2020-04-26T18:30:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.