Cross-Forgery Analysis of Vision Transformers and CNNs for Deepfake
Image Detection
- URL: http://arxiv.org/abs/2206.13829v1
- Date: Tue, 28 Jun 2022 08:50:22 GMT
- Title: Cross-Forgery Analysis of Vision Transformers and CNNs for Deepfake
Image Detection
- Authors: Davide Alessandro Coccomini, Roberto Caldelli, Fabrizio Falchi,
Claudio Gennaro, Giuseppe Amato
- Abstract summary: We show that EfficientNetV2 has a greater tendency to specialize often obtaining better results on training methods.
We also show that Vision Transformers exhibit a superior generalization ability that makes them more competent even on images generated with new methodologies.
- Score: 11.944111906027144
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deepfake Generation Techniques are evolving at a rapid pace, making it
possible to create realistic manipulated images and videos and endangering the
serenity of modern society. The continual emergence of new and varied
techniques brings with it a further problem to be faced, namely the ability of
deepfake detection models to update themselves promptly in order to be able to
identify manipulations carried out using even the most recent methods. This is
an extremely complex problem to solve, as training a model requires large
amounts of data, which are difficult to obtain if the deepfake generation
method is too recent. Moreover, continuously retraining a network would be
unfeasible. In this paper, we ask ourselves if, among the various deep learning
techniques, there is one that is able to generalise the concept of deepfake to
such an extent that it does not remain tied to one or more specific deepfake
generation methods used in the training set. We compared a Vision Transformer
with an EfficientNetV2 on a cross-forgery context based on the ForgeryNet
dataset. From our experiments, It emerges that EfficientNetV2 has a greater
tendency to specialize often obtaining better results on training methods while
Vision Transformers exhibit a superior generalization ability that makes them
more competent even on images generated with new methodologies.
Related papers
- Segue: Side-information Guided Generative Unlearnable Examples for
Facial Privacy Protection in Real World [64.4289385463226]
We propose Segue: Side-information guided generative unlearnable examples.
To improve transferability, we introduce side information such as true labels and pseudo labels.
It can resist JPEG compression, adversarial training, and some standard data augmentations.
arXiv Detail & Related papers (2023-10-24T06:22:37Z) - Towards General Visual-Linguistic Face Forgery Detection [95.73987327101143]
Deepfakes are realistic face manipulations that can pose serious threats to security, privacy, and trust.
Existing methods mostly treat this task as binary classification, which uses digital labels or mask signals to train the detection model.
We propose a novel paradigm named Visual-Linguistic Face Forgery Detection(VLFFD), which uses fine-grained sentence-level prompts as the annotation.
arXiv Detail & Related papers (2023-07-31T10:22:33Z) - Deep Image Fingerprint: Towards Low Budget Synthetic Image Detection and Model Lineage Analysis [8.777277201807351]
We develop a new detection method for images that are indistinguishable from real ones.
Our method can detect images from a known generative model and enable us to establish relationships between fine-tuned generative models.
Our approach achieves comparable performance to state-of-the-art pre-trained detection methods on images generated by Stable Diffusion and Midversa.
arXiv Detail & Related papers (2023-03-19T20:31:38Z) - Deep Convolutional Pooling Transformer for Deepfake Detection [54.10864860009834]
We propose a deep convolutional Transformer to incorporate decisive image features both locally and globally.
Specifically, we apply convolutional pooling and re-attention to enrich the extracted features and enhance efficacy.
The proposed solution consistently outperforms several state-of-the-art baselines on both within- and cross-dataset experiments.
arXiv Detail & Related papers (2022-09-12T15:05:41Z) - Hybrid Transformer Network for Deepfake Detection [2.644723682054489]
We propose a novel hybrid transformer network utilizing early feature fusion strategy for deepfake video detection.
Our model achieves comparable results to other more advanced state-of-the-art approaches when evaluated on FaceForensics++ and DFDC benchmarks.
We also propose novel face cut-out augmentations, as well as random cut-out augmentations.
arXiv Detail & Related papers (2022-08-11T13:30:42Z) - Self-supervised Transformer for Deepfake Detection [112.81127845409002]
Deepfake techniques in real-world scenarios require stronger generalization abilities of face forgery detectors.
Inspired by transfer learning, neural networks pre-trained on other large-scale face-related tasks may provide useful features for deepfake detection.
In this paper, we propose a self-supervised transformer based audio-visual contrastive learning method.
arXiv Detail & Related papers (2022-03-02T17:44:40Z) - FReTAL: Generalizing Deepfake Detection using Knowledge Distillation and
Representation Learning [17.97648576135166]
We introduce a transfer learning-based Feature Representation Transfer Adaptation Learning (FReTAL) method.
Our student model can quickly adapt to new types of deepfake by distilling knowledge from a pre-trained teacher model.
FReTAL outperforms all baselines on the domain adaptation task with up to 86.97% accuracy on low-quality deepfakes.
arXiv Detail & Related papers (2021-05-28T06:54:10Z) - Deepfake Forensics via An Adversarial Game [99.84099103679816]
We advocate adversarial training for improving the generalization ability to both unseen facial forgeries and unseen image/video qualities.
Considering that AI-based face manipulation often leads to high-frequency artifacts that can be easily spotted by models yet difficult to generalize, we propose a new adversarial training method that attempts to blur out these specific artifacts.
arXiv Detail & Related papers (2021-03-25T02:20:08Z) - Artificial Fingerprinting for Generative Models: Rooting Deepfake
Attribution in Training Data [64.65952078807086]
Photorealistic image generation has reached a new level of quality due to the breakthroughs of generative adversarial networks (GANs)
Yet, the dark side of such deepfakes, the malicious use of generated media, raises concerns about visual misinformation.
We seek a proactive and sustainable solution on deepfake detection by introducing artificial fingerprints into the models.
arXiv Detail & Related papers (2020-07-16T16:49:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.