MD-CSDNetwork: Multi-Domain Cross Stitched Network for Deepfake
Detection
- URL: http://arxiv.org/abs/2109.07311v1
- Date: Wed, 15 Sep 2021 14:11:53 GMT
- Title: MD-CSDNetwork: Multi-Domain Cross Stitched Network for Deepfake
Detection
- Authors: Aayushi Agarwal, Akshay Agarwal, Sayan Sinha, Mayank Vatsa, Richa
Singh
- Abstract summary: Current deepfake generation methods leave discriminative artifacts in the frequency spectrum of fake images and videos.
We present a novel approach, termed as MD-CSDNetwork, for combining the features in the spatial and frequency domains to mine a shared discriminative representation.
- Score: 80.83725644958633
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The rapid progress in the ease of creating and spreading ultra-realistic
media over social platforms calls for an urgent need to develop a generalizable
deepfake detection technique. It has been observed that current deepfake
generation methods leave discriminative artifacts in the frequency spectrum of
fake images and videos. Inspired by this observation, in this paper, we present
a novel approach, termed as MD-CSDNetwork, for combining the features in the
spatial and frequency domains to mine a shared discriminative representation
for classifying \textit{deepfakes}. MD-CSDNetwork is a novel cross-stitched
network with two parallel branches carrying the spatial and frequency
information, respectively. We hypothesize that these multi-domain input data
streams can be considered as related supervisory signals. The supervision from
both branches ensures better performance and generalization. Further, the
concept of cross-stitch connections is utilized where they are inserted between
the two branches to learn an optimal combination of domain-specific and shared
representations from other domains automatically. Extensive experiments are
conducted on the popular benchmark dataset namely FaceForeniscs++ for forgery
classification. We report improvements over all the manipulation types in
FaceForensics++ dataset and comparable results with state-of-the-art methods
for cross-database evaluation on the Celeb-DF dataset and the Deepfake
Detection Dataset.
Related papers
- Contextual Cross-Modal Attention for Audio-Visual Deepfake Detection and Localization [3.9440964696313485]
In the digital age, the emergence of deepfakes and synthetic media presents a significant threat to societal and political integrity.
Deepfakes based on multi-modal manipulation, such as audio-visual, are more realistic and pose a greater threat.
We propose a novel multi-modal attention framework based on recurrent neural networks (RNNs) that leverages contextual information for audio-visual deepfake detection.
arXiv Detail & Related papers (2024-08-02T18:45:01Z) - Deep Homography Estimation for Visual Place Recognition [49.235432979736395]
We propose a transformer-based deep homography estimation (DHE) network.
It takes the dense feature map extracted by a backbone network as input and fits homography for fast and learnable geometric verification.
Experiments on benchmark datasets show that our method can outperform several state-of-the-art methods.
arXiv Detail & Related papers (2024-02-25T13:22:17Z) - ATTA: Anomaly-aware Test-Time Adaptation for Out-of-Distribution
Detection in Segmentation [22.084967085509387]
We propose a dual-level OOD detection framework to handle domain shift and semantic shift jointly.
The first level distinguishes whether domain shift exists in the image by leveraging global low-level features.
The second level identifies pixels with semantic shift by utilizing dense high-level feature maps.
arXiv Detail & Related papers (2023-09-12T06:49:56Z) - Hierarchical Forgery Classifier On Multi-modality Face Forgery Clues [61.37306431455152]
We propose a novel Hierarchical Forgery for Multi-modality Face Forgery Detection (HFC-MFFD)
The HFC-MFFD learns robust patches-based hybrid representation to enhance forgery authentication in multiple-modality scenarios.
The specific hierarchical face forgery is proposed to alleviate the class imbalance problem and further boost detection performance.
arXiv Detail & Related papers (2022-12-30T10:54:29Z) - Deepfake Detection via Joint Unsupervised Reconstruction and Supervised
Classification [25.84902508816679]
We introduce a novel approach for deepfake detection, which considers the reconstruction and classification tasks simultaneously.
This method shares the information learned by one task with the other, which focuses on a different aspect other existing works rarely consider.
Our method achieves state-of-the-art performance on three commonly-used datasets.
arXiv Detail & Related papers (2022-11-24T05:44:26Z) - Cross-Domain Facial Expression Recognition: A Unified Evaluation
Benchmark and Adversarial Graph Learning [85.6386289476598]
We develop a novel adversarial graph representation adaptation (AGRA) framework for cross-domain holistic-local feature co-adaptation.
We conduct extensive and fair evaluations on several popular benchmarks and show that the proposed AGRA framework outperforms previous state-of-the-art methods.
arXiv Detail & Related papers (2020-08-03T15:00:31Z) - Cross-layer Feature Pyramid Network for Salient Object Detection [102.20031050972429]
We propose a novel Cross-layer Feature Pyramid Network to improve the progressive fusion in salient object detection.
The distributed features per layer own both semantics and salient details from all other layers simultaneously, and suffer reduced loss of important information.
arXiv Detail & Related papers (2020-02-25T14:06:27Z) - Universal-RCNN: Universal Object Detector via Transferable Graph R-CNN [117.80737222754306]
We present a novel universal object detector called Universal-RCNN.
We first generate a global semantic pool by integrating all high-level semantic representation of all the categories.
An Intra-Domain Reasoning Module learns and propagates the sparse graph representation within one dataset guided by a spatial-aware GCN.
arXiv Detail & Related papers (2020-02-18T07:57:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.