SLIP: Spoof-Aware One-Class Face Anti-Spoofing with Language Image Pretraining
- URL: http://arxiv.org/abs/2503.19982v1
- Date: Tue, 25 Mar 2025 18:09:06 GMT
- Title: SLIP: Spoof-Aware One-Class Face Anti-Spoofing with Language Image Pretraining
- Authors: Pei-Kai Huang, Jun-Xiong Chong, Cheng-Hsuan Chiang, Tzu-Hsien Chen, Tyng-Luh Liu, Chiou-Ting Hsu,
- Abstract summary: Face anti-spoofing (FAS) plays a pivotal role in ensuring the security and reliability of face recognition systems.<n>One-class FAS focuses on learning intrinsic liveness features solely from live training images to differentiate between live and spoof faces.<n>We propose a novel framework called Spoof-aware one-class face anti-spoofing with Language Image Pretraining (SLIP)
- Score: 12.144338080864127
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Face anti-spoofing (FAS) plays a pivotal role in ensuring the security and reliability of face recognition systems. With advancements in vision-language pretrained (VLP) models, recent two-class FAS techniques have leveraged the advantages of using VLP guidance, while this potential remains unexplored in one-class FAS methods. The one-class FAS focuses on learning intrinsic liveness features solely from live training images to differentiate between live and spoof faces. However, the lack of spoof training data can lead one-class FAS models to inadvertently incorporate domain information irrelevant to the live/spoof distinction (e.g., facial content), causing performance degradation when tested with a new application domain. To address this issue, we propose a novel framework called Spoof-aware one-class face anti-spoofing with Language Image Pretraining (SLIP). Given that live faces should ideally not be obscured by any spoof-attack-related objects (e.g., paper, or masks) and are assumed to yield zero spoof cue maps, we first propose an effective language-guided spoof cue map estimation to enhance one-class FAS models by simulating whether the underlying faces are covered by attack-related objects and generating corresponding nonzero spoof cue maps. Next, we introduce a novel prompt-driven liveness feature disentanglement to alleviate live/spoof-irrelative domain variations by disentangling live/spoof-relevant and domain-dependent information. Finally, we design an effective augmentation strategy by fusing latent features from live images and spoof prompts to generate spoof-like image features and thus diversify latent spoof features to facilitate the learning of one-class FAS. Our extensive experiments and ablation studies support that SLIP consistently outperforms previous one-class FAS methods.
Related papers
- Unsupervised Feature Disentanglement and Augmentation Network for One-class Face Anti-spoofing [3.5032444938182477]
Face anti-sfing (FAS) techniques aim to enhance the security of facial identity authentication by distinguishing authentic live faces from deceptive attempts.
One-class FAS approaches handle unseen attacks well but are less robust to domain information entangled within the liveness features.
We propose an Unsupervised Feature Disentanglement and Augmentation Network (textbfUFDANet) that enhances generalizability by augmenting face images via disentangled features.
arXiv Detail & Related papers (2025-03-29T01:22:50Z) - Interpretable Face Anti-Spoofing: Enhancing Generalization with Multimodal Large Language Models [58.936893810674896]
Face Anti-Spoofing (FAS) is essential for ensuring the security and reliability of facial recognition systems.<n>We introduce a multimodal large language model framework for FAS, termed Interpretable Face Anti-Spoofing (I-FAS)<n>We propose a Spoof-aware Captioning and Filtering (SCF) strategy to generate high-quality captions for FAS images.
arXiv Detail & Related papers (2025-01-03T09:25:04Z) - DiffFAS: Face Anti-Spoofing via Generative Diffusion Models [27.533334690705733]
Face anti-spoofing (FAS) plays a vital role in preventing face recognition (FR) systems from presentation attacks.
We propose DiffFAS framework, which quantifies quality as prior information input into the network to counter image quality shift.
We demonstrate the effectiveness of our framework on challenging cross-domain and cross-attack FAS datasets.
arXiv Detail & Related papers (2024-09-13T06:45:23Z) - UniForensics: Face Forgery Detection via General Facial Representation [60.5421627990707]
High-level semantic features are less susceptible to perturbations and not limited to forgery-specific artifacts, thus having stronger generalization.
We introduce UniForensics, a novel deepfake detection framework that leverages a transformer-based video network, with a meta-functional face classification for enriched facial representation.
arXiv Detail & Related papers (2024-07-26T20:51:54Z) - Attribute-preserving Face Dataset Anonymization via Latent Code
Optimization [64.4569739006591]
We present a task-agnostic anonymization procedure that directly optimize the images' latent representation in the latent space of a pre-trained GAN.
We demonstrate through a series of experiments that our method is capable of anonymizing the identity of the images whilst -- crucially -- better-preserving the facial attributes.
arXiv Detail & Related papers (2023-03-20T17:34:05Z) - Learning Facial Liveness Representation for Domain Generalized Face
Anti-spoofing [25.07432145233952]
Face anti-spoofing (FAS) aims at distinguishing face spoof attacks from the authentic ones.
It is not practical to assume that the type of spoof attacks would be known in advance.
We propose a deep learning model for addressing the aforementioned domain-generalized face anti-spoofing task.
arXiv Detail & Related papers (2022-08-16T16:13:24Z) - PatchNet: A Simple Face Anti-Spoofing Framework via Fine-Grained Patch
Recognition [13.840830140721462]
Face anti-spoofing (FAS) plays a critical role in securing face recognition systems from presentation attacks.
We propose PatchNet which reformulates face anti-spoofing as a fine-grained patch-type recognition problem.
arXiv Detail & Related papers (2022-03-27T15:16:17Z) - Dual Spoof Disentanglement Generation for Face Anti-spoofing with Depth
Uncertainty Learning [54.15303628138665]
Face anti-spoofing (FAS) plays a vital role in preventing face recognition systems from presentation attacks.
Existing face anti-spoofing datasets lack diversity due to the insufficient identity and insignificant variance.
We propose Dual Spoof Disentanglement Generation framework to tackle this challenge by "anti-spoofing via generation"
arXiv Detail & Related papers (2021-12-01T15:36:59Z) - Disentangled Representation with Dual-stage Feature Learning for Face
Anti-spoofing [18.545438302664756]
It is essential to learn more generalized and discriminative features to prevent overfitting to pre-defined spoof attack types.
This paper proposes a novel dual-stage disentangled representation learning method that can efficiently untangle spoof-related features from irrelevant ones.
arXiv Detail & Related papers (2021-10-18T10:22:52Z) - Uncertainty-Aware Physically-Guided Proxy Tasks for Unseen Domain Face
Anti-spoofing [128.32381246318954]
Face anti-spoofing (FAS) seeks to discriminate genuine faces from fake ones arising from any type of spoofing attack.
We propose to leverage physical cues to attain better generalization on unseen domains.
arXiv Detail & Related papers (2020-11-28T03:22:26Z) - Face Anti-Spoofing Via Disentangled Representation Learning [90.90512800361742]
Face anti-spoofing is crucial to security of face recognition systems.
We propose a novel perspective of face anti-spoofing that disentangles the liveness features and content features from images.
arXiv Detail & Related papers (2020-08-19T03:54:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.