Physically-Based Face Rendering for NIR-VIS Face Recognition
- URL: http://arxiv.org/abs/2211.06408v1
- Date: Fri, 11 Nov 2022 18:48:16 GMT
- Title: Physically-Based Face Rendering for NIR-VIS Face Recognition
- Authors: Yunqi Miao, Alexandros Lattas, Jiankang Deng, Jungong Han, Stefanos
Zafeiriou
- Abstract summary: Near infrared (NIR) to Visible (VIS) face matching is challenging due to the significant domain gaps.
We propose a novel method for paired NIR-VIS facial image generation.
To facilitate the identity feature learning, we propose an IDentity-based Maximum Mean Discrepancy (ID-MMD) loss.
- Score: 165.54414962403555
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Near infrared (NIR) to Visible (VIS) face matching is challenging due to the
significant domain gaps as well as a lack of sufficient data for cross-modality
model training. To overcome this problem, we propose a novel method for paired
NIR-VIS facial image generation. Specifically, we reconstruct 3D face shape and
reflectance from a large 2D facial dataset and introduce a novel method of
transforming the VIS reflectance to NIR reflectance. We then use a
physically-based renderer to generate a vast, high-resolution and
photorealistic dataset consisting of various poses and identities in the NIR
and VIS spectra. Moreover, to facilitate the identity feature learning, we
propose an IDentity-based Maximum Mean Discrepancy (ID-MMD) loss, which not
only reduces the modality gap between NIR and VIS images at the domain level
but encourages the network to focus on the identity features instead of facial
details, such as poses and accessories. Extensive experiments conducted on four
challenging NIR-VIS face recognition benchmarks demonstrate that the proposed
method can achieve comparable performance with the state-of-the-art (SOTA)
methods without requiring any existing NIR-VIS face recognition datasets. With
slightly fine-tuning on the target NIR-VIS face recognition datasets, our
method can significantly surpass the SOTA performance. Code and pretrained
models are released under the insightface
(https://github.com/deepinsight/insightface/tree/master/recognition).
Related papers
- Improving Neural Surface Reconstruction with Feature Priors from Multi-View Image [87.00660347447494]
Recent advancements in Neural Surface Reconstruction (NSR) have significantly improved multi-view reconstruction when coupled with volume rendering.
We propose an investigation into feature-level consistent loss, aiming to harness valuable feature priors from diverse pretext visual tasks.
Our results, analyzed on DTU and EPFL, reveal that feature priors from image matching and multi-view stereo datasets outperform other pretext tasks.
arXiv Detail & Related papers (2024-08-04T16:09:46Z) - Hypergraph-Guided Disentangled Spectrum Transformer Networks for
Near-Infrared Facial Expression Recognition [31.783671943393344]
We give the first attempt to deep NIR facial expression recognition and proposed a novel method called near-infrared facial expression transformer (NFER-Former)
NFER-Former disentangles the expression information and spectrum information from the input image, so that the expression features can be extracted without the interference of spectrum variation.
We have constructed a large NIR-VIS Facial Expression dataset that includes 360 subjects to better validate the efficiency of NFER-Former.
arXiv Detail & Related papers (2023-12-10T15:15:50Z) - Improving 2D face recognition via fine-level facial depth generation and
RGB-D complementary feature learning [0.8223798883838329]
We propose a fine-grained facial depth generation network and an improved multimodal complementary feature learning network.
Experiments on the Lock3DFace dataset and the IIIT-D dataset show that the proposed FFDGNet and I MCFLNet can improve the accuracy of RGB-D face recognition.
arXiv Detail & Related papers (2023-05-08T02:33:59Z) - Mutimodal Ranking Optimization for Heterogeneous Face Re-identification [17.987122302811404]
Heterogeneous face re-identification, namely matching heterogeneous faces across disjoint visible light (VIS) and near-infrared (NIR) cameras, has become an important problem in video surveillance application.
To solve this problem, a multimodal fusion ranking optimization algorithm for heterogeneous face re-identification is proposed in this paper.
arXiv Detail & Related papers (2022-12-11T14:04:24Z) - Towards NIR-VIS Masked Face Recognition [47.00916333095693]
Near-infrared to visible (NIR-VIS) face recognition is the most common case in heterogeneous face recognition.
We propose a novel training method to maximize the mutual information shared by the face representation of two domains.
In addition, a 3D face reconstruction based approach is employed to synthesize masked face from the existing NIR image.
arXiv Detail & Related papers (2021-04-14T10:40:09Z) - Hierarchical Deep CNN Feature Set-Based Representation Learning for
Robust Cross-Resolution Face Recognition [59.29808528182607]
Cross-resolution face recognition (CRFR) is important in intelligent surveillance and biometric forensics.
Existing shallow learning-based and deep learning-based methods focus on mapping the HR-LR face pairs into a joint feature space.
In this study, we desire to fully exploit the multi-level deep convolutional neural network (CNN) feature set for robust CRFR.
arXiv Detail & Related papers (2021-03-25T14:03:42Z) - A NIR-to-VIS face recognition via part adaptive and relation attention
module [4.822208985805956]
In the face recognition application scenario, we need to process facial images captured in various conditions, such as at night by near-infrared (NIR) surveillance cameras.
The illumination difference between NIR and visible-light (VIS) causes a domain gap between facial images, and the variations in pose and emotion also make facial matching more difficult.
We propose a part relation attention module that crops facial parts obtained through a semantic mask and performs relational modeling using each of these representative features.
arXiv Detail & Related papers (2021-02-01T08:13:39Z) - The FaceChannel: A Fast & Furious Deep Neural Network for Facial
Expression Recognition [71.24825724518847]
Current state-of-the-art models for automatic Facial Expression Recognition (FER) are based on very deep neural networks that are effective but rather expensive to train.
We formalize the FaceChannel, a light-weight neural network that has much fewer parameters than common deep neural networks.
We demonstrate how our model achieves a comparable, if not better, performance to the current state-of-the-art in FER.
arXiv Detail & Related papers (2020-09-15T09:25:37Z) - DotFAN: A Domain-transferred Face Augmentation Network for Pose and
Illumination Invariant Face Recognition [94.96686189033869]
We propose a 3D model-assisted domain-transferred face augmentation network (DotFAN)
DotFAN can generate a series of variants of an input face based on the knowledge distilled from existing rich face datasets collected from other domains.
Experiments show that DotFAN is beneficial for augmenting small face datasets to improve their within-class diversity.
arXiv Detail & Related papers (2020-02-23T08:16:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.