A NIR-to-VIS face recognition via part adaptive and relation attention
module
- URL: http://arxiv.org/abs/2102.00689v1
- Date: Mon, 1 Feb 2021 08:13:39 GMT
- Title: A NIR-to-VIS face recognition via part adaptive and relation attention
module
- Authors: Rushuang Xu, MyeongAh Cho, and Sangyoun Lee
- Abstract summary: In the face recognition application scenario, we need to process facial images captured in various conditions, such as at night by near-infrared (NIR) surveillance cameras.
The illumination difference between NIR and visible-light (VIS) causes a domain gap between facial images, and the variations in pose and emotion also make facial matching more difficult.
We propose a part relation attention module that crops facial parts obtained through a semantic mask and performs relational modeling using each of these representative features.
- Score: 4.822208985805956
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the face recognition application scenario, we need to process facial
images captured in various conditions, such as at night by near-infrared (NIR)
surveillance cameras. The illumination difference between NIR and visible-light
(VIS) causes a domain gap between facial images, and the variations in pose and
emotion also make facial matching more difficult. Heterogeneous face
recognition (HFR) has difficulties in domain discrepancy, and many studies have
focused on extracting domain-invariant features, such as facial part relational
information. However, when pose variation occurs, the facial component position
changes, and a different part relation is extracted. In this paper, we propose
a part relation attention module that crops facial parts obtained through a
semantic mask and performs relational modeling using each of these
representative features. Furthermore, we suggest component adaptive triplet
loss function using adaptive weights for each part to reduce the intra-class
identity regardless of the domain as well as pose. Finally, our method exhibits
a performance improvement in the CASIA NIR-VIS 2.0 and achieves superior result
in the BUAA-VisNir with large pose and emotion variations.
Related papers
- Text-Guided Face Recognition using Multi-Granularity Cross-Modal
Contrastive Learning [0.0]
We introduce text-guided face recognition (TGFR) to analyze the impact of integrating facial attributes in the form of natural language descriptions.
TGFR demonstrates remarkable improvements, particularly on low-quality images, over existing face recognition models.
arXiv Detail & Related papers (2023-12-14T22:04:22Z) - Physically-Based Face Rendering for NIR-VIS Face Recognition [165.54414962403555]
Near infrared (NIR) to Visible (VIS) face matching is challenging due to the significant domain gaps.
We propose a novel method for paired NIR-VIS facial image generation.
To facilitate the identity feature learning, we propose an IDentity-based Maximum Mean Discrepancy (ID-MMD) loss.
arXiv Detail & Related papers (2022-11-11T18:48:16Z) - FaceDancer: Pose- and Occlusion-Aware High Fidelity Face Swapping [62.38898610210771]
We present a new single-stage method for subject face swapping and identity transfer, named FaceDancer.
We have two major contributions: Adaptive Feature Fusion Attention (AFFA) and Interpreted Feature Similarity Regularization (IFSR)
arXiv Detail & Related papers (2022-10-19T11:31:38Z) - CIAO! A Contrastive Adaptation Mechanism for Non-Universal Facial
Expression Recognition [80.07590100872548]
We propose Contrastive Inhibitory Adaptati On (CIAO), a mechanism that adapts the last layer of facial encoders to depict specific affective characteristics on different datasets.
CIAO presents an improvement in facial expression recognition performance over six different datasets with very unique affective representations.
arXiv Detail & Related papers (2022-08-10T15:46:05Z) - NIR-to-VIS Face Recognition via Embedding Relations and Coordinates of
the Pairwise Features [5.044100238869375]
We propose a 'Relation Module' which can simply add-on to any face recognition models.
The local features extracted from face image contain information of each component of the face.
With the proposed module, we achieve 14.81% rank-1 accuracy and 15.47% verification rate of 0.1% FAR improvements.
arXiv Detail & Related papers (2022-08-04T02:53:44Z) - PAM: Pose Attention Module for Pose-Invariant Face Recognition [3.0839245814393723]
We propose a lightweight and easy-to-implement attention block, named Pose Attention Module (PAM), for pose-invariant face recognition.
Specifically, PAM performs frontal-profile feature transformation in hierarchical feature space by learning residuals between pose variations with a soft gate mechanism.
arXiv Detail & Related papers (2021-11-23T15:18:33Z) - Towards NIR-VIS Masked Face Recognition [47.00916333095693]
Near-infrared to visible (NIR-VIS) face recognition is the most common case in heterogeneous face recognition.
We propose a novel training method to maximize the mutual information shared by the face representation of two domains.
In addition, a 3D face reconstruction based approach is employed to synthesize masked face from the existing NIR image.
arXiv Detail & Related papers (2021-04-14T10:40:09Z) - Cross-Domain Identification for Thermal-to-Visible Face Recognition [6.224425156703344]
This paper proposes a novel domain adaptation framework that combines a new feature mapping sub-network with existing deep feature models.
New cross-domain identity and domain invariance loss functions for thermal-to-visible face recognition alleviates the requirement for precisely co-registered and synchronized imagery.
We analyze the viability of the proposed framework for more challenging tasks, such as non-frontal thermal-to-visible face recognition.
arXiv Detail & Related papers (2020-08-19T14:24:04Z) - InterFaceGAN: Interpreting the Disentangled Face Representation Learned
by GANs [73.27299786083424]
We propose a framework called InterFaceGAN to interpret the disentangled face representation learned by state-of-the-art GAN models.
We first find that GANs learn various semantics in some linear subspaces of the latent space.
We then conduct a detailed study on the correlation between different semantics and manage to better disentangle them via subspace projection.
arXiv Detail & Related papers (2020-05-18T18:01:22Z) - DotFAN: A Domain-transferred Face Augmentation Network for Pose and
Illumination Invariant Face Recognition [94.96686189033869]
We propose a 3D model-assisted domain-transferred face augmentation network (DotFAN)
DotFAN can generate a series of variants of an input face based on the knowledge distilled from existing rich face datasets collected from other domains.
Experiments show that DotFAN is beneficial for augmenting small face datasets to improve their within-class diversity.
arXiv Detail & Related papers (2020-02-23T08:16:34Z) - Dual-Attention GAN for Large-Pose Face Frontalization [59.689836951934694]
We present a novel Dual-Attention Generative Adversarial Network (DA-GAN) for photo-realistic face frontalization.
Specifically, a self-attention-based generator is introduced to integrate local features with their long-range dependencies.
A novel face-attention-based discriminator is applied to emphasize local features of face regions.
arXiv Detail & Related papers (2020-02-17T20:00:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.