Two-branch Recurrent Network for Isolating Deepfakes in Videos
- URL: http://arxiv.org/abs/2008.03412v3
- Date: Fri, 4 Sep 2020 01:03:55 GMT
- Title: Two-branch Recurrent Network for Isolating Deepfakes in Videos
- Authors: Iacopo Masi, Aditya Killekar, Royston Marian Mascarenhas, Shenoy
Pratik Gurudatt, Wael AbdAlmageed
- Abstract summary: We present a method for deepfake detection based on a two-branch network structure.
One branch propagates the original information, while the other branch suppresses the face content.
Our two novel components show promising results on the FaceForensics++, Celeb-DF, and Facebook's DFDC preview benchmarks.
- Score: 17.59209853264258
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The current spike of hyper-realistic faces artificially generated using
deepfakes calls for media forensics solutions that are tailored to video
streams and work reliably with a low false alarm rate at the video level. We
present a method for deepfake detection based on a two-branch network structure
that isolates digitally manipulated faces by learning to amplify artifacts
while suppressing the high-level face content. Unlike current methods that
extract spatial frequencies as a preprocessing step, we propose a two-branch
structure: one branch propagates the original information, while the other
branch suppresses the face content yet amplifies multi-band frequencies using a
Laplacian of Gaussian (LoG) as a bottleneck layer. To better isolate
manipulated faces, we derive a novel cost function that, unlike regular
classification, compresses the variability of natural faces and pushes away the
unrealistic facial samples in the feature space. Our two novel components show
promising results on the FaceForensics++, Celeb-DF, and Facebook's DFDC preview
benchmarks, when compared to prior work. We then offer a full, detailed
ablation study of our network architecture and cost function. Finally, although
the bar is still high to get very remarkable figures at a very low false alarm
rate, our study shows that we can achieve good video-level performance when
cross-testing in terms of video-level AUC.
Related papers
- UniForensics: Face Forgery Detection via General Facial Representation [60.5421627990707]
High-level semantic features are less susceptible to perturbations and not limited to forgery-specific artifacts, thus having stronger generalization.
We introduce UniForensics, a novel deepfake detection framework that leverages a transformer-based video network, with a meta-functional face classification for enriched facial representation.
arXiv Detail & Related papers (2024-07-26T20:51:54Z) - AltFreezing for More General Video Face Forgery Detection [138.5732617371004]
We propose to capture both spatial and unseen temporal artifacts in one model for face forgery detection.
We present a novel training strategy called AltFreezing for more general face forgery detection.
arXiv Detail & Related papers (2023-07-17T08:24:58Z) - Unmasking Deepfakes: Masked Autoencoding Spatiotemporal Transformers for
Enhanced Video Forgery Detection [19.432851794777754]
We present a novel approach for the detection of deepfake videos using a pair of vision transformers pre-trained by a self-supervised masked autoencoding setup.
Our method consists of two distinct components, one of which focuses on learning spatial information from individual RGB frames of the video, while the other learns temporal consistency information from optical flow fields generated from consecutive frames.
arXiv Detail & Related papers (2023-06-12T05:49:23Z) - Deep Convolutional Pooling Transformer for Deepfake Detection [54.10864860009834]
We propose a deep convolutional Transformer to incorporate decisive image features both locally and globally.
Specifically, we apply convolutional pooling and re-attention to enrich the extracted features and enhance efficacy.
The proposed solution consistently outperforms several state-of-the-art baselines on both within- and cross-dataset experiments.
arXiv Detail & Related papers (2022-09-12T15:05:41Z) - Spatial-Temporal Frequency Forgery Clue for Video Forgery Detection in
VIS and NIR Scenario [87.72258480670627]
Existing face forgery detection methods based on frequency domain find that the GAN forged images have obvious grid-like visual artifacts in the frequency spectrum compared to the real images.
This paper proposes a Cosine Transform-based Forgery Clue Augmentation Network (FCAN-DCT) to achieve a more comprehensive spatial-temporal feature representation.
arXiv Detail & Related papers (2022-07-05T09:27:53Z) - Voice-Face Homogeneity Tells Deepfake [56.334968246631725]
Existing detection approaches contribute to exploring the specific artifacts in deepfake videos.
We propose to perform the deepfake detection from an unexplored voice-face matching view.
Our model obtains significantly improved performance as compared to other state-of-the-art competitors.
arXiv Detail & Related papers (2022-03-04T09:08:50Z) - Sharp Multiple Instance Learning for DeepFake Video Detection [54.12548421282696]
We introduce a new problem of partial face attack in DeepFake video, where only video-level labels are provided but not all the faces in the fake videos are manipulated.
A sharp MIL (S-MIL) is proposed which builds direct mapping from instance embeddings to bag prediction.
Experiments on FFPMS and widely used DFDC dataset verify that S-MIL is superior to other counterparts for partially attacked DeepFake video detection.
arXiv Detail & Related papers (2020-08-11T08:52:17Z) - Deep Frequent Spatial Temporal Learning for Face Anti-Spoofing [9.435020319411311]
Face anti-spoofing is crucial for the security of face recognition system, by avoiding invaded with presentation attack.
Previous works have shown the effectiveness of using depth and temporal supervision for this task.
We propose a novel two stream FreqSaptialTemporalNet for face anti-spoofing which simultaneously takes advantage of frequent, spatial and temporal information.
arXiv Detail & Related papers (2020-01-20T06:02:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.