AltFreezing for More General Video Face Forgery Detection
- URL: http://arxiv.org/abs/2307.08317v1
- Date: Mon, 17 Jul 2023 08:24:58 GMT
- Title: AltFreezing for More General Video Face Forgery Detection
- Authors: Zhendong Wang, Jianmin Bao, Wengang Zhou, Weilun Wang, Houqiang Li
- Abstract summary: We propose to capture both spatial and unseen temporal artifacts in one model for face forgery detection.
We present a novel training strategy called AltFreezing for more general face forgery detection.
- Score: 138.5732617371004
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Existing face forgery detection models try to discriminate fake images by
detecting only spatial artifacts (e.g., generative artifacts, blending) or
mainly temporal artifacts (e.g., flickering, discontinuity). They may
experience significant performance degradation when facing out-domain
artifacts. In this paper, we propose to capture both spatial and temporal
artifacts in one model for face forgery detection. A simple idea is to leverage
a spatiotemporal model (3D ConvNet). However, we find that it may easily rely
on one type of artifact and ignore the other. To address this issue, we present
a novel training strategy called AltFreezing for more general face forgery
detection. The AltFreezing aims to encourage the model to detect both spatial
and temporal artifacts. It divides the weights of a spatiotemporal network into
two groups: spatial-related and temporal-related. Then the two groups of
weights are alternately frozen during the training process so that the model
can learn spatial and temporal features to distinguish real or fake videos.
Furthermore, we introduce various video-level data augmentation methods to
improve the generalization capability of the forgery detection model. Extensive
experiments show that our framework outperforms existing methods in terms of
generalization to unseen manipulations and datasets. Code is available at
https: //github.com/ZhendongWang6/AltFreezing.
Related papers
- Generalizing Deepfake Video Detection with Plug-and-Play: Video-Level Blending and Spatiotemporal Adapter Tuning [42.86270268974854]
Temporal features can be complex and diverse.
Spatiotemporal models often lean heavily on one type of artifact and ignore the other.
Videos are naturally resource-intensive.
arXiv Detail & Related papers (2024-08-30T07:49:57Z) - UniForensics: Face Forgery Detection via General Facial Representation [60.5421627990707]
High-level semantic features are less susceptible to perturbations and not limited to forgery-specific artifacts, thus having stronger generalization.
We introduce UniForensics, a novel deepfake detection framework that leverages a transformer-based video network, with a meta-functional face classification for enriched facial representation.
arXiv Detail & Related papers (2024-07-26T20:51:54Z) - SIGMA:Sinkhorn-Guided Masked Video Modeling [69.31715194419091]
Sinkhorn-guided Masked Video Modelling ( SIGMA) is a novel video pretraining method.
We distribute features of space-time tubes evenly across a limited number of learnable clusters.
Experimental results on ten datasets validate the effectiveness of SIGMA in learning more performant, temporally-aware, and robust video representations.
arXiv Detail & Related papers (2024-07-22T08:04:09Z) - Latent Spatiotemporal Adaptation for Generalized Face Forgery Video Detection [22.536129731902783]
We propose a Latemporal Spatio(LAST) approach to facilitate generalized face video detection.
We first model thetemporal patterns face videos by incorporating a lightweight CNN to extract local spatial features of each frame.
Then we learn the long-termtemporal representations in latent space videos, which should contain more clues than in pixel space.
arXiv Detail & Related papers (2023-09-09T13:40:44Z) - Detection of Deepfake Videos Using Long Distance Attention [73.6659488380372]
Most existing detection methods treat the problem as a vanilla binary classification problem.
In this paper, the problem is treated as a special fine-grained classification problem since the differences between fake and real faces are very subtle.
A spatial-temporal model is proposed which has two components for capturing spatial and temporal forgery traces in global perspective.
arXiv Detail & Related papers (2021-06-24T08:33:32Z) - ST-HOI: A Spatial-Temporal Baseline for Human-Object Interaction
Detection in Videos [91.29436920371003]
We propose a simple yet effective architecture named Spatial-Temporal HOI Detection (ST-HOI)
We use temporal information such as human and object trajectories, correctly-localized visual features, and spatial-temporal masking pose features.
We construct a new video HOI benchmark dubbed VidHOI where our proposed approach serves as a solid baseline.
arXiv Detail & Related papers (2021-05-25T07:54:35Z) - CutPaste: Self-Supervised Learning for Anomaly Detection and
Localization [59.719925639875036]
We propose a framework for building anomaly detectors using normal training data only.
We first learn self-supervised deep representations and then build a generative one-class classifier on learned representations.
Our empirical study on MVTec anomaly detection dataset demonstrates the proposed algorithm is general to be able to detect various types of real-world defects.
arXiv Detail & Related papers (2021-04-08T19:04:55Z) - Two-branch Recurrent Network for Isolating Deepfakes in Videos [17.59209853264258]
We present a method for deepfake detection based on a two-branch network structure.
One branch propagates the original information, while the other branch suppresses the face content.
Our two novel components show promising results on the FaceForensics++, Celeb-DF, and Facebook's DFDC preview benchmarks.
arXiv Detail & Related papers (2020-08-08T01:38:56Z) - Deepfake Detection using Spatiotemporal Convolutional Networks [0.0]
Deepfake detection methods only use individual frames and therefore fail to learn from temporal information.
We created a benchmark of performance using Celeb-DF dataset.
Our methods outperformed state-of-theart frame-based detection methods.
arXiv Detail & Related papers (2020-06-26T01:32:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.