Joint Feature Distribution Alignment Learning for NIR-VIS and VIS-VIS
Face Recognition
- URL: http://arxiv.org/abs/2204.11434v1
- Date: Mon, 25 Apr 2022 04:47:35 GMT
- Title: Joint Feature Distribution Alignment Learning for NIR-VIS and VIS-VIS
Face Recognition
- Authors: Takaya Miyamoto, Hiroshi Hashimoto, Akihiro Hayasaka, Akinori F.
Ebihara, Hitoshi Imaoka
- Abstract summary: heterogeneous face recognition (HFR) is still a difficult task due to the domain discrepancy and lack of large HFR dataset.
We propose joint feature distribution alignment learning (JFDAL) which is a joint learning approach utilizing knowledge distillation.
Our method achieves a comparable HFR performance on the Oulu-CASIA NIR&VIS dataset with less degradation of VIS performance.
- Score: 5.249805590164902
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Face recognition for visible light (VIS) images achieve high accuracy thanks
to the recent development of deep learning. However, heterogeneous face
recognition (HFR), which is a face matching in different domains, is still a
difficult task due to the domain discrepancy and lack of large HFR dataset.
Several methods have attempted to reduce the domain discrepancy by means of
fine-tuning, which causes significant degradation of the performance in the VIS
domain because it loses the highly discriminative VIS representation. To
overcome this problem, we propose joint feature distribution alignment learning
(JFDAL) which is a joint learning approach utilizing knowledge distillation. It
enables us to achieve high HFR performance with retaining the original
performance for the VIS domain. Extensive experiments demonstrate that our
proposed method delivers statistically significantly better performances
compared with the conventional fine-tuning approach on a public HFR dataset
Oulu-CASIA NIR&VIS and popular verification datasets in VIS domain such as FLW,
CFP, AgeDB. Furthermore, comparative experiments with existing state-of-the-art
HFR methods show that our method achieves a comparable HFR performance on the
Oulu-CASIA NIR&VIS dataset with less degradation of VIS performance.
Related papers
- Towards Data-Centric Face Anti-Spoofing: Improving Cross-domain Generalization via Physics-based Data Synthesis [64.46312434121455]
Face Anti-Spoofing (FAS) research is challenged by the cross-domain problem, where there is a domain gap between the training and testing data.
We propose task-specific FAS data augmentation (FAS-Aug), which increases data diversity by synthesizing data of artifacts.
We also propose Spoofing Attack Risk Equalization (SARE) to prevent models from relying on certain types of artifacts and improve the generalization performance.
arXiv Detail & Related papers (2024-09-04T01:45:18Z) - Heterogeneous Face Recognition Using Domain Invariant Units [4.910937238451485]
We leverage a pretrained face recognition model as a teacher network to learn domaininvariant network layers called Domain-Invariant Units (DIU)
The proposed DIU can be trained effectively even with a limited amount of paired training data, in a contrastive distillation framework.
This proposed approach has the potential to enhance pretrained models, making them more adaptable to a wider range of variations in data.
arXiv Detail & Related papers (2024-04-22T16:58:37Z) - CATFace: Cross-Attribute-Guided Transformer with Self-Attention
Distillation for Low-Quality Face Recognition [12.129404936688752]
We propose a novel multi-branch neural network that leverages soft biometric (SB) attribute information to boost the performance of face recognition (FR)
We introduce a novel self-attention distillation framework that effectively highlights crucial facial regions, such as landmarks by aligning low-quality images with those of their high-quality counterparts in the feature space.
arXiv Detail & Related papers (2024-01-05T19:21:31Z) - Rethinking the Domain Gap in Near-infrared Face Recognition [65.7871950460781]
Heterogeneous face recognition (HFR) involves the intricate task of matching face images across the visual domains of visible (VIS) and near-infrared (NIR)
Much of the existing literature on HFR identifies the domain gap as a primary challenge and directs efforts towards bridging it at either the input or feature level.
We observe that large neural networks, unlike their smaller counterparts, when pre-trained on large scale homogeneous VIS data, demonstrate exceptional zero-shot performance in HFR.
arXiv Detail & Related papers (2023-12-01T14:43:28Z) - Physically-Based Face Rendering for NIR-VIS Face Recognition [165.54414962403555]
Near infrared (NIR) to Visible (VIS) face matching is challenging due to the significant domain gaps.
We propose a novel method for paired NIR-VIS facial image generation.
To facilitate the identity feature learning, we propose an IDentity-based Maximum Mean Discrepancy (ID-MMD) loss.
arXiv Detail & Related papers (2022-11-11T18:48:16Z) - Cluster-level pseudo-labelling for source-free cross-domain facial
expression recognition [94.56304526014875]
We propose the first Source-Free Unsupervised Domain Adaptation (SFUDA) method for Facial Expression Recognition (FER)
Our method exploits self-supervised pretraining to learn good feature representations from the target data.
We validate the effectiveness of our method in four adaptation setups, proving that it consistently outperforms existing SFUDA methods when applied to FER.
arXiv Detail & Related papers (2022-10-11T08:24:50Z) - Escaping Data Scarcity for High-Resolution Heterogeneous Face
Hallucination [68.78903256687697]
In Heterogeneous Face Recognition (HFR), the objective is to match faces across two different domains such as visible and thermal.
Recent methods attempting to fill the gap via synthesis have achieved promising results, but their performance is still limited by the scarcity of paired training data.
In this paper, we propose a new face hallucination paradigm for HFR, which not only enables data-efficient synthesis but also allows to scale up model training without breaking any privacy policy.
arXiv Detail & Related papers (2022-03-30T20:44:33Z) - On Exploring Pose Estimation as an Auxiliary Learning Task for
Visible-Infrared Person Re-identification [66.58450185833479]
In this paper, we exploit Pose Estimation as an auxiliary learning task to assist the VI-ReID task in an end-to-end framework.
By jointly training these two tasks in a mutually beneficial manner, our model learns higher quality modality-shared and ID-related features.
Experimental results on two benchmark VI-ReID datasets show that the proposed method consistently improves state-of-the-art methods by significant margins.
arXiv Detail & Related papers (2022-01-11T09:44:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.