End2End Occluded Face Recognition by Masking Corrupted Features
- URL: http://arxiv.org/abs/2108.09468v1
- Date: Sat, 21 Aug 2021 09:08:41 GMT
- Title: End2End Occluded Face Recognition by Masking Corrupted Features
- Authors: Haibo Qiu, Dihong Gong, Zhifeng Li, Wei Liu, Dacheng Tao
- Abstract summary: State-of-the-art general face recognition models do not generalize well to occluded face images.
This paper presents a novel face recognition method that is robust to occlusions based on a single end-to-end deep neural network.
Our approach, named FROM (Face Recognition with Occlusion Masks), learns to discover the corrupted features from the deep convolutional neural networks, and clean them by the dynamically learned masks.
- Score: 82.27588990277192
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: With the recent advancement of deep convolutional neural networks,
significant progress has been made in general face recognition. However, the
state-of-the-art general face recognition models do not generalize well to
occluded face images, which are exactly the common cases in real-world
scenarios. The potential reasons are the absences of large-scale occluded face
data for training and specific designs for tackling corrupted features brought
by occlusions. This paper presents a novel face recognition method that is
robust to occlusions based on a single end-to-end deep neural network. Our
approach, named FROM (Face Recognition with Occlusion Masks), learns to
discover the corrupted features from the deep convolutional neural networks,
and clean them by the dynamically learned masks. In addition, we construct
massive occluded face images to train FROM effectively and efficiently. FROM is
simple yet powerful compared to the existing methods that either rely on
external detectors to discover the occlusions or employ shallow models which
are less discriminative. Experimental results on the LFW, Megaface challenge 1,
RMF2, AR dataset and other simulated occluded/masked datasets confirm that FROM
dramatically improves the accuracy under occlusions, and generalizes well on
general face recognition.
Related papers
- Seeing through the Mask: Multi-task Generative Mask Decoupling Face
Recognition [47.248075664420874]
Current general face recognition system suffers from serious performance degradation when encountering occluded scenes.
This paper proposes a Multi-task gEnerative mask dEcoupling face Recognition (MEER) network to jointly handle these two tasks.
We first present a novel mask decoupling module to disentangle mask and identity information, which makes the network obtain purer identity features from visible facial components.
arXiv Detail & Related papers (2023-11-20T03:23:03Z) - Latent-OFER: Detect, Mask, and Reconstruct with Latent Vectors for
Occluded Facial Expression Recognition [0.0]
The proposed method can detect occluded parts of the face as if they were unoccluded, and recognize them, improving FER accuracy.
It involves three steps: First, the vision transformer (ViT)-based occlusion patch detector masks the occluded position by training only latent vectors from the unoccluded patches.
Second, the hybrid reconstruction network generates the masking position as a complete image using the ViT and convolutional neural network (CNN)
Last, the expression-relevant latent vector extractor retrieves and uses expression-related information from all latent vectors by applying a CNN-based class activation map
arXiv Detail & Related papers (2023-07-21T07:56:32Z) - Real Face Foundation Representation Learning for Generalized Deepfake
Detection [74.4691295738097]
The emergence of deepfake technologies has become a matter of social concern as they pose threats to individual privacy and public security.
It is almost impossible to collect sufficient representative fake faces, and it is hard for existing detectors to generalize to all types of manipulation.
We propose Real Face Foundation Representation Learning (RFFR), which aims to learn a general representation from large-scale real face datasets.
arXiv Detail & Related papers (2023-03-15T08:27:56Z) - FaceOcc: A Diverse, High-quality Face Occlusion Dataset for Human Face
Extraction [3.8502825594372703]
Occlusions often occur in face images in the wild, troubling face-related tasks such as landmark detection, 3D reconstruction, and face recognition.
This paper proposes a novel face segmentation dataset with manually labeled face occlusions from the CelebA-HQ and the internet.
We trained a straightforward face segmentation model but obtained SOTA performance, convincingly demonstrating the effectiveness of the proposed dataset.
arXiv Detail & Related papers (2022-01-20T19:44:18Z) - Towards NIR-VIS Masked Face Recognition [47.00916333095693]
Near-infrared to visible (NIR-VIS) face recognition is the most common case in heterogeneous face recognition.
We propose a novel training method to maximize the mutual information shared by the face representation of two domains.
In addition, a 3D face reconstruction based approach is employed to synthesize masked face from the existing NIR image.
arXiv Detail & Related papers (2021-04-14T10:40:09Z) - When Face Recognition Meets Occlusion: A New Benchmark [37.616211206620854]
We create a simulated occlusion face recognition dataset.
It covers 804,704 face images of 10,575 subjects.
Our dataset significantly outperforms the state-of-the-arts.
arXiv Detail & Related papers (2021-03-04T03:07:42Z) - Face Hallucination via Split-Attention in Split-Attention Network [58.30436379218425]
convolutional neural networks (CNNs) have been widely employed to promote the face hallucination.
We propose a novel external-internal split attention group (ESAG) to take into account the overall facial profile and fine texture details simultaneously.
By fusing the features from these two paths, the consistency of facial structure and the fidelity of facial details are strengthened.
arXiv Detail & Related papers (2020-10-22T10:09:31Z) - Exploiting Semantics for Face Image Deblurring [121.44928934662063]
We propose an effective and efficient face deblurring algorithm by exploiting semantic cues via deep convolutional neural networks.
We incorporate face semantic labels as input priors and propose an adaptive structural loss to regularize facial local structures.
The proposed method restores sharp images with more accurate facial features and details.
arXiv Detail & Related papers (2020-01-19T13:06:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.