Fake Visual Content Detection Using Two-Stream Convolutional Neural
Networks
- URL: http://arxiv.org/abs/2101.00676v1
- Date: Sun, 3 Jan 2021 18:05:07 GMT
- Title: Fake Visual Content Detection Using Two-Stream Convolutional Neural
Networks
- Authors: Bilal Yousaf, Muhammad Usama, Waqas Sultani, Arif Mahmood, Junaid
Qadir
- Abstract summary: We propose a two-stream convolutional neural network architecture called TwoStreamNet to complement frequency and spatial domain features.
The proposed detector has demonstrated significant performance improvement compared to the current state-of-the-art fake content detectors.
- Score: 14.781702606707642
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Rapid progress in adversarial learning has enabled the generation of
realistic-looking fake visual content. To distinguish between fake and real
visual content, several detection techniques have been proposed. The
performance of most of these techniques however drops off significantly if the
test and the training data are sampled from different distributions. This
motivates efforts towards improving the generalization of fake detectors. Since
current fake content generation techniques do not accurately model the
frequency spectrum of the natural images, we observe that the frequency
spectrum of the fake visual data contains discriminative characteristics that
can be used to detect fake content. We also observe that the information
captured in the frequency spectrum is different from that of the spatial
domain. Using these insights, we propose to complement frequency and spatial
domain features using a two-stream convolutional neural network architecture
called TwoStreamNet. We demonstrate the improved generalization of the proposed
two-stream network to several unseen generation architectures, datasets, and
techniques. The proposed detector has demonstrated significant performance
improvement compared to the current state-of-the-art fake content detectors and
fusing the frequency and spatial domain streams has also improved
generalization of the detector.
Related papers
- Detecting Inpainted Video with Frequency Domain Insights [0.0]
We propose the Frequency Domain Insights Network (FDIN), which significantly enhances detection accuracy.
Previous evaluations on public datasets demonstrate that FDIN achieves state-of-the-art performance.
arXiv Detail & Related papers (2024-09-21T01:51:07Z) - Multiple Contexts and Frequencies Aggregation Network forDeepfake Detection [5.65128683992597]
Deepfake detection faces increasing challenges since the fast growth of generative models in developing massive and diverse Deepfake technologies.
Recent advances rely on introducing features from spatial or frequency domains rather than modeling general forgery features within backbones.
We propose an efficient network for face forgery detection named MkfaNet, which consists of two core modules.
arXiv Detail & Related papers (2024-08-03T05:34:53Z) - Frequency-Aware Deepfake Detection: Improving Generalizability through
Frequency Space Learning [81.98675881423131]
This research addresses the challenge of developing a universal deepfake detector that can effectively identify unseen deepfake images.
Existing frequency-based paradigms have relied on frequency-level artifacts introduced during the up-sampling in GAN pipelines to detect forgeries.
We introduce a novel frequency-aware approach called FreqNet, centered around frequency domain learning, specifically designed to enhance the generalizability of deepfake detectors.
arXiv Detail & Related papers (2024-03-12T01:28:00Z) - Multimodal Graph Learning for Deepfake Detection [10.077496841634135]
Existing deepfake detectors face several challenges in achieving robustness and generalization.
We propose a novel framework, namely Multimodal Graph Learning (MGL), that leverages information from multiple modalities.
Our proposed method aims to effectively identify and utilize distinguishing features for deepfake detection.
arXiv Detail & Related papers (2022-09-12T17:17:49Z) - Deep Convolutional Pooling Transformer for Deepfake Detection [54.10864860009834]
We propose a deep convolutional Transformer to incorporate decisive image features both locally and globally.
Specifically, we apply convolutional pooling and re-attention to enrich the extracted features and enhance efficacy.
The proposed solution consistently outperforms several state-of-the-art baselines on both within- and cross-dataset experiments.
arXiv Detail & Related papers (2022-09-12T15:05:41Z) - Spatial-Temporal Frequency Forgery Clue for Video Forgery Detection in
VIS and NIR Scenario [87.72258480670627]
Existing face forgery detection methods based on frequency domain find that the GAN forged images have obvious grid-like visual artifacts in the frequency spectrum compared to the real images.
This paper proposes a Cosine Transform-based Forgery Clue Augmentation Network (FCAN-DCT) to achieve a more comprehensive spatial-temporal feature representation.
arXiv Detail & Related papers (2022-07-05T09:27:53Z) - Adaptive Frequency Learning in Two-branch Face Forgery Detection [66.91715092251258]
We propose Adaptively learn Frequency information in the two-branch Detection framework, dubbed AFD.
We liberate our network from the fixed frequency transforms, and achieve better performance with our data- and task-dependent transform layers.
arXiv Detail & Related papers (2022-03-27T14:25:52Z) - Attention-based Dual-stream Vision Transformer for Radar Gait
Recognition [24.90100456414406]
Radar gait recognition is robust to light variations and less infringement on privacy.
In this work, a dual-stream neural network with attention-based fusion is proposed to fully aggregate the discriminant information.
The proposed method is validated on a large benchmark dataset for radar gait recognition, which shows that it significantly outperforms state-of-the-art solutions.
arXiv Detail & Related papers (2021-11-24T06:16:53Z) - MD-CSDNetwork: Multi-Domain Cross Stitched Network for Deepfake
Detection [80.83725644958633]
Current deepfake generation methods leave discriminative artifacts in the frequency spectrum of fake images and videos.
We present a novel approach, termed as MD-CSDNetwork, for combining the features in the spatial and frequency domains to mine a shared discriminative representation.
arXiv Detail & Related papers (2021-09-15T14:11:53Z) - Beyond the Spectrum: Detecting Deepfakes via Re-Synthesis [69.09526348527203]
Deep generative models have led to highly realistic media, known as deepfakes, that are commonly indistinguishable from real to human eyes.
We propose a novel fake detection that is designed to re-synthesize testing images and extract visual cues for detection.
We demonstrate the improved effectiveness, cross-GAN generalization, and robustness against perturbations of our approach in a variety of detection scenarios.
arXiv Detail & Related papers (2021-05-29T21:22:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.