Interpretable and Trustworthy Deepfake Detection via Dynamic Prototypes
- URL: http://arxiv.org/abs/2006.15473v2
- Date: Fri, 15 Jan 2021 02:13:45 GMT
- Title: Interpretable and Trustworthy Deepfake Detection via Dynamic Prototypes
- Authors: Loc Trinh, Michael Tsang, Sirisha Rambhatla, Yan Liu
- Abstract summary: We propose a novel human-centered approach for detecting forgery in face images, using dynamic prototypes as a form of visual explanations.
Extensive experimental results show that DPNet achieves competitive predictive performance, even on unseen testing datasets.
- Score: 20.358053429294458
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper we propose a novel human-centered approach for detecting
forgery in face images, using dynamic prototypes as a form of visual
explanations. Currently, most state-of-the-art deepfake detections are based on
black-box models that process videos frame-by-frame for inference, and few
closely examine their temporal inconsistencies. However, the existence of such
temporal artifacts within deepfake videos is key in detecting and explaining
deepfakes to a supervising human. To this end, we propose Dynamic Prototype
Network (DPNet) -- an interpretable and effective solution that utilizes
dynamic representations (i.e., prototypes) to explain deepfake temporal
artifacts. Extensive experimental results show that DPNet achieves competitive
predictive performance, even on unseen testing datasets such as Google's
DeepFakeDetection, DeeperForensics, and Celeb-DF, while providing easy
referential explanations of deepfake dynamics. On top of DPNet's prototypical
framework, we further formulate temporal logic specifications based on these
dynamics to check our model's compliance to desired temporal behaviors, hence
providing trustworthiness for such critical detection systems.
Related papers
- FakeFormer: Efficient Vulnerability-Driven Transformers for Generalisable Deepfake Detection [12.594436202557446]
This paper investigates why Vision Transformers (ViTs) exhibit a suboptimal performance when dealing with the detection of facial forgeries.
We propose a deepfake detection framework called FakeFormer, which extends ViTs to enforce the extraction of subtle inconsistency-prone information.
Experiments are conducted on diverse well-known datasets, including FF++, Celeb-DF, WildDeepfake, DFD, DFDCP, and DFDC.
arXiv Detail & Related papers (2024-10-29T11:36:49Z) - UniForensics: Face Forgery Detection via General Facial Representation [60.5421627990707]
High-level semantic features are less susceptible to perturbations and not limited to forgery-specific artifacts, thus having stronger generalization.
We introduce UniForensics, a novel deepfake detection framework that leverages a transformer-based video network, with a meta-functional face classification for enriched facial representation.
arXiv Detail & Related papers (2024-07-26T20:51:54Z) - Rethinking the Up-Sampling Operations in CNN-based Generative Network
for Generalizable Deepfake Detection [86.97062579515833]
We introduce the concept of Neighboring Pixel Relationships(NPR) as a means to capture and characterize the generalized structural artifacts stemming from up-sampling operations.
A comprehensive analysis is conducted on an open-world dataset, comprising samples generated by tft28 distinct generative models.
This analysis culminates in the establishment of a novel state-of-the-art performance, showcasing a remarkable tft11.6% improvement over existing methods.
arXiv Detail & Related papers (2023-12-16T14:27:06Z) - ProtoExplorer: Interpretable Forensic Analysis of Deepfake Videos using
Prototype Exploration and Refinement [11.182863992851622]
ProtoExplorer is a Visual Analytics system for exploration and refinement of prototype-based deepfake detection models.
It offers tools for visualizing and temporally filtering prototype-based predictions when working with video data.
System was designed with forensic experts and evaluated in a number of rounds based on both open-ended think evaluation and interviews.
arXiv Detail & Related papers (2023-09-20T09:03:56Z) - Towards General Visual-Linguistic Face Forgery Detection [95.73987327101143]
Deepfakes are realistic face manipulations that can pose serious threats to security, privacy, and trust.
Existing methods mostly treat this task as binary classification, which uses digital labels or mask signals to train the detection model.
We propose a novel paradigm named Visual-Linguistic Face Forgery Detection(VLFFD), which uses fine-grained sentence-level prompts as the annotation.
arXiv Detail & Related papers (2023-07-31T10:22:33Z) - Deep Convolutional Pooling Transformer for Deepfake Detection [54.10864860009834]
We propose a deep convolutional Transformer to incorporate decisive image features both locally and globally.
Specifically, we apply convolutional pooling and re-attention to enrich the extracted features and enhance efficacy.
The proposed solution consistently outperforms several state-of-the-art baselines on both within- and cross-dataset experiments.
arXiv Detail & Related papers (2022-09-12T15:05:41Z) - The Effectiveness of Temporal Dependency in Deepfake Video Detection [0.0]
This paper investigates whether temporal information can improve the deepfake performance of deep learning models.
We find that temporal dependency produces a statistically significant increase in performance classifying real images for the model.
arXiv Detail & Related papers (2022-05-13T14:39:25Z) - Beyond the Spectrum: Detecting Deepfakes via Re-Synthesis [69.09526348527203]
Deep generative models have led to highly realistic media, known as deepfakes, that are commonly indistinguishable from real to human eyes.
We propose a novel fake detection that is designed to re-synthesize testing images and extract visual cues for detection.
We demonstrate the improved effectiveness, cross-GAN generalization, and robustness against perturbations of our approach in a variety of detection scenarios.
arXiv Detail & Related papers (2021-05-29T21:22:24Z) - DS-Net: Dynamic Spatiotemporal Network for Video Salient Object
Detection [78.04869214450963]
We propose a novel dynamic temporal-temporal network (DSNet) for more effective fusion of temporal and spatial information.
We show that the proposed method achieves superior performance than state-of-the-art algorithms.
arXiv Detail & Related papers (2020-12-09T06:42:30Z) - Spatio-temporal Features for Generalized Detection of Deepfake Videos [12.453288832098314]
We propose-temporal features, modeled by 3D CNNs, to extend the capabilities to detect new sorts of deep videos.
We show that our approach outperforms existing methods in terms of generalization capabilities.
arXiv Detail & Related papers (2020-10-22T16:28:50Z) - Dynamic texture analysis for detecting fake faces in video sequences [6.1356022122903235]
This work explores the analysis of texture-temporal dynamics of the video signal.
The goal is to characterizing and distinguishing real fake sequences.
We propose to build multiple binary decision on the joint analysis of temporal segments.
arXiv Detail & Related papers (2020-07-30T07:21:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.