Mitigating Adversarial Attacks in Deepfake Detection: An Exploration of
Perturbation and AI Techniques
- URL: http://arxiv.org/abs/2302.11704v2
- Date: Sun, 10 Sep 2023 00:22:22 GMT
- Title: Mitigating Adversarial Attacks in Deepfake Detection: An Exploration of
Perturbation and AI Techniques
- Authors: Saminder Dhesi, Laura Fontes, Pedro Machado, Isibor Kennedy Ihianle,
Farhad Fassihi Tash, David Ada Adama
- Abstract summary: adversarial examples are subtle perturbations artfully injected into clean images or videos.
Deepfakes have emerged as a potent tool to manipulate public opinion and tarnish the reputations of public figures.
This article delves into the multifaceted world of adversarial examples, elucidating the underlying principles behind their capacity to deceive deep learning algorithms.
- Score: 1.0718756132502771
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Deep learning constitutes a pivotal component within the realm of machine
learning, offering remarkable capabilities in tasks ranging from image
recognition to natural language processing. However, this very strength also
renders deep learning models susceptible to adversarial examples, a phenomenon
pervasive across a diverse array of applications. These adversarial examples
are characterized by subtle perturbations artfully injected into clean images
or videos, thereby causing deep learning algorithms to misclassify or produce
erroneous outputs. This susceptibility extends beyond the confines of digital
domains, as adversarial examples can also be strategically designed to target
human cognition, leading to the creation of deceptive media, such as deepfakes.
Deepfakes, in particular, have emerged as a potent tool to manipulate public
opinion and tarnish the reputations of public figures, underscoring the urgent
need to address the security and ethical implications associated with
adversarial examples. This article delves into the multifaceted world of
adversarial examples, elucidating the underlying principles behind their
capacity to deceive deep learning algorithms. We explore the various
manifestations of this phenomenon, from their insidious role in compromising
model reliability to their impact in shaping the contemporary landscape of
disinformation and misinformation. To illustrate progress in combating
adversarial examples, we showcase the development of a tailored Convolutional
Neural Network (CNN) designed explicitly to detect deepfakes, a pivotal step
towards enhancing model robustness in the face of adversarial threats.
Impressively, this custom CNN has achieved a precision rate of 76.2% on the
DFDC dataset.
Related papers
- A Survey on Transferability of Adversarial Examples across Deep Neural Networks [53.04734042366312]
adversarial examples can manipulate machine learning models into making erroneous predictions.
The transferability of adversarial examples enables black-box attacks which circumvent the need for detailed knowledge of the target model.
This survey explores the landscape of the adversarial transferability of adversarial examples.
arXiv Detail & Related papers (2023-10-26T17:45:26Z) - Investigating Human-Identifiable Features Hidden in Adversarial
Perturbations [54.39726653562144]
Our study explores up to five attack algorithms across three datasets.
We identify human-identifiable features in adversarial perturbations.
Using pixel-level annotations, we extract such features and demonstrate their ability to compromise target models.
arXiv Detail & Related papers (2023-09-28T22:31:29Z) - Deviations in Representations Induced by Adversarial Attacks [0.0]
Research has shown that deep learning models are vulnerable to adversarial attacks.
This finding brought about a new direction in research, whereby algorithms were developed to attack and defend vulnerable networks.
We present a method for measuring and analyzing the deviations in representations induced by adversarial attacks.
arXiv Detail & Related papers (2022-11-07T17:40:08Z) - Attack to Fool and Explain Deep Networks [59.97135687719244]
We counter-argue by providing evidence of human-meaningful patterns in adversarial perturbations.
Our major contribution is a novel pragmatic adversarial attack that is subsequently transformed into a tool to interpret the visual models.
arXiv Detail & Related papers (2021-06-20T03:07:36Z) - Explainable Adversarial Attacks in Deep Neural Networks Using Activation
Profiles [69.9674326582747]
This paper presents a visual framework to investigate neural network models subjected to adversarial examples.
We show how observing these elements can quickly pinpoint exploited areas in a model.
arXiv Detail & Related papers (2021-03-18T13:04:21Z) - Detection Defense Against Adversarial Attacks with Saliency Map [7.736844355705379]
It is well established that neural networks are vulnerable to adversarial examples, which are almost imperceptible on human vision.
Existing defenses are trend to harden the robustness of models against adversarial attacks.
We propose a novel method combined with additional noises and utilize the inconsistency strategy to detect adversarial examples.
arXiv Detail & Related papers (2020-09-06T13:57:17Z) - Towards Transferable Adversarial Attack against Deep Face Recognition [58.07786010689529]
Deep convolutional neural networks (DCNNs) have been found to be vulnerable to adversarial examples.
transferable adversarial examples can severely hinder the robustness of DCNNs.
We propose DFANet, a dropout-based method used in convolutional layers, which can increase the diversity of surrogate models.
We generate a new set of adversarial face pairs that can successfully attack four commercial APIs without any queries.
arXiv Detail & Related papers (2020-04-13T06:44:33Z) - Plausible Counterfactuals: Auditing Deep Learning Classifiers with
Realistic Adversarial Examples [84.8370546614042]
Black-box nature of Deep Learning models has posed unanswered questions about what they learn from data.
Generative Adversarial Network (GAN) and multi-objectives are used to furnish a plausible attack to the audited model.
Its utility is showcased within a human face classification task, unveiling the enormous potential of the proposed framework.
arXiv Detail & Related papers (2020-03-25T11:08:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.