Creating Artificial Modalities to Solve RGB Liveness
- URL: http://arxiv.org/abs/2006.16028v1
- Date: Mon, 29 Jun 2020 13:19:22 GMT
- Title: Creating Artificial Modalities to Solve RGB Liveness
- Authors: Aleksandr Parkin and Oleg Grinchuk
- Abstract summary: We introduce two types of artificial transforms: rank pooling and optical flow, combined in end-to-end pipeline for spoof detection.
The proposed method achieves state-of-the-art on the largest cross-ethnicity face anti-spoofing dataset CASIA-SURF CeFA (RGB)
- Score: 79.9255035557979
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Special cameras that provide useful features for face anti-spoofing are
desirable, but not always an option. In this work we propose a method to
utilize the difference in dynamic appearance between bona fide and spoof
samples by creating artificial modalities from RGB videos. We introduce two
types of artificial transforms: rank pooling and optical flow, combined in
end-to-end pipeline for spoof detection. We demonstrate that using intermediate
representations that contain less identity and fine-grained features increase
model robustness to unseen attacks as well as to unseen ethnicities. The
proposed method achieves state-of-the-art on the largest cross-ethnicity face
anti-spoofing dataset CASIA-SURF CeFA (RGB).
Related papers
- Adv-Diffusion: Imperceptible Adversarial Face Identity Attack via Latent
Diffusion Model [61.53213964333474]
We propose a unified framework Adv-Diffusion that can generate imperceptible adversarial identity perturbations in the latent space but not the raw pixel space.
Specifically, we propose the identity-sensitive conditioned diffusion generative model to generate semantic perturbations in the surroundings.
The designed adaptive strength-based adversarial perturbation algorithm can ensure both attack transferability and stealthiness.
arXiv Detail & Related papers (2023-12-18T15:25:23Z) - RANRAC: Robust Neural Scene Representations via Random Ray Consensus [12.161889666145127]
RANdom RAy Consensus (RANRAC) is an efficient approach to eliminate the effect of inconsistent data.
We formulate a fuzzy adaption of the RANSAC paradigm, enabling its application to large scale models.
Results indicate significant improvements compared to state-of-the-art robust methods for novel-view synthesis.
arXiv Detail & Related papers (2023-12-15T13:33:09Z) - Attentive Multimodal Fusion for Optical and Scene Flow [24.08052492109655]
Existing methods typically rely solely on RGB images or fuse the modalities at later stages.
We propose a novel deep neural network approach named FusionRAFT, which enables early-stage information fusion between sensor modalities.
Our approach exhibits improved robustness in the presence of noise and low-lighting conditions that affect the RGB images.
arXiv Detail & Related papers (2023-07-28T04:36:07Z) - A Multi-modal Approach to Single-modal Visual Place Classification [2.580765958706854]
Multi-sensor fusion approaches combining RGB and depth (D) have gained popularity in recent years.
We reformulate the single-modal RGB image classification task as a pseudo multi-modal RGB-D classification problem.
A practical, fully self-supervised framework for training, appropriately processing, fusing, and classifying these two modalities is described.
arXiv Detail & Related papers (2023-05-10T14:04:21Z) - Complementary Random Masking for RGB-Thermal Semantic Segmentation [63.93784265195356]
RGB-thermal semantic segmentation is a potential solution to achieve reliable semantic scene understanding in adverse weather and lighting conditions.
This paper proposes 1) a complementary random masking strategy of RGB-T images and 2) self-distillation loss between clean and masked input modalities.
We achieve state-of-the-art performance over three RGB-T semantic segmentation benchmarks.
arXiv Detail & Related papers (2023-03-30T13:57:21Z) - Multi-Modal Human Authentication Using Silhouettes, Gait and RGB [59.46083527510924]
Whole-body-based human authentication is a promising approach for remote biometrics scenarios.
We propose Dual-Modal Ensemble (DME), which combines both RGB and silhouette data to achieve more robust performances for indoor and outdoor whole-body based recognition.
Within DME, we propose GaitPattern, which is inspired by the double helical gait pattern used in traditional gait analysis.
arXiv Detail & Related papers (2022-10-08T15:17:32Z) - Dual Spoof Disentanglement Generation for Face Anti-spoofing with Depth
Uncertainty Learning [54.15303628138665]
Face anti-spoofing (FAS) plays a vital role in preventing face recognition systems from presentation attacks.
Existing face anti-spoofing datasets lack diversity due to the insufficient identity and insignificant variance.
We propose Dual Spoof Disentanglement Generation framework to tackle this challenge by "anti-spoofing via generation"
arXiv Detail & Related papers (2021-12-01T15:36:59Z) - Two-stream Convolutional Networks for Multi-frame Face Anti-spoofing [1.9890930069402575]
We propose an efficient two-stream model to capture the key differences between live and spoof faces.
We evaluate the proposed method on the datasets of Siw, Oulu-NPU, CASIA-MFSD and Replay-Attack.
arXiv Detail & Related papers (2021-08-09T13:35:30Z) - Camera Invariant Feature Learning for Generalized Face Anti-spoofing [95.30490139294136]
We describe a framework that eliminates the influence of inherent variance from acquisition cameras at the feature level.
Experiments show that the proposed method can achieve better performance in both intra-dataset and cross-dataset settings.
arXiv Detail & Related papers (2021-01-25T13:40:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.