Fairness Evaluation in Deepfake Detection Models using Metamorphic
Testing
- URL: http://arxiv.org/abs/2203.06825v1
- Date: Mon, 14 Mar 2022 02:44:56 GMT
- Title: Fairness Evaluation in Deepfake Detection Models using Metamorphic
Testing
- Authors: Muxin Pu, Meng Yi Kuan, Nyee Thoang Lim, Chun Yong Chong, Mei Kuan Lim
- Abstract summary: This work is to evaluate how deepfake detection model behaves under such anomalies.
We have chosen MesoInception-4, a state-of-the-art deepfake detection model, as the target model.
We focus on revealing potential gender biases in DL and AI systems.
- Score: 2.0649235321315285
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Fairness of deepfake detectors in the presence of anomalies are not well
investigated, especially if those anomalies are more prominent in either male
or female subjects. The primary motivation for this work is to evaluate how
deepfake detection model behaves under such anomalies. However, due to the
black-box nature of deep learning (DL) and artificial intelligence (AI)
systems, it is hard to predict the performance of a model when the input data
is modified. Crucially, if this defect is not addressed properly, it will
adversely affect the fairness of the model and result in discrimination of
certain sub-population unintentionally. Therefore, the objective of this work
is to adopt metamorphic testing to examine the reliability of the selected
deepfake detection model, and how the transformation of input variation places
influence on the output. We have chosen MesoInception-4, a state-of-the-art
deepfake detection model, as the target model and makeup as the anomalies.
Makeups are applied through utilizing the Dlib library to obtain the 68 facial
landmarks prior to filling in the RGB values. Metamorphic relations are derived
based on the notion that realistic perturbations of the input images, such as
makeup, involving eyeliners, eyeshadows, blushes, and lipsticks (which are
common cosmetic appearance) applied to male and female images, should not alter
the output of the model by a huge margin. Furthermore, we narrow down the scope
to focus on revealing potential gender biases in DL and AI systems.
Specifically, we are interested to examine whether MesoInception-4 model
produces unfair decisions, which should be considered as a consequence of
robustness issues. The findings from our work have the potential to pave the
way for new research directions in the quality assurance and fairness in DL and
AI systems.
Related papers
- MeLIAD: Interpretable Few-Shot Anomaly Detection with Metric Learning and Entropy-based Scoring [2.394081903745099]
We propose MeLIAD, a novel methodology for interpretable anomaly detection.
MeLIAD is based on metric learning and achieves interpretability by design without relying on any prior distribution assumptions of true anomalies.
Experiments on five public benchmark datasets, including quantitative and qualitative evaluation of interpretability, demonstrate that MeLIAD achieves improved anomaly detection and localization performance.
arXiv Detail & Related papers (2024-09-20T16:01:43Z) - Self-supervised Feature Adaptation for 3D Industrial Anomaly Detection [59.41026558455904]
We focus on multi-modal anomaly detection. Specifically, we investigate early multi-modal approaches that attempted to utilize models pre-trained on large-scale visual datasets.
We propose a Local-to-global Self-supervised Feature Adaptation (LSFA) method to finetune the adaptors and learn task-oriented representation toward anomaly detection.
arXiv Detail & Related papers (2024-01-06T07:30:41Z) - AnomalyDiffusion: Few-Shot Anomaly Image Generation with Diffusion Model [59.08735812631131]
Anomaly inspection plays an important role in industrial manufacture.
Existing anomaly inspection methods are limited in their performance due to insufficient anomaly data.
We propose AnomalyDiffusion, a novel diffusion-based few-shot anomaly generation model.
arXiv Detail & Related papers (2023-12-10T05:13:40Z) - Unleashing Mask: Explore the Intrinsic Out-of-Distribution Detection
Capability [70.72426887518517]
Out-of-distribution (OOD) detection is an indispensable aspect of secure AI when deploying machine learning models in real-world applications.
We propose a novel method, Unleashing Mask, which aims to restore the OOD discriminative capabilities of the well-trained model with ID data.
Our method utilizes a mask to figure out the memorized atypical samples, and then finetune the model or prune it with the introduced mask to forget them.
arXiv Detail & Related papers (2023-06-06T14:23:34Z) - Fast Unsupervised Brain Anomaly Detection and Segmentation with
Diffusion Models [1.6352599467675781]
We propose a method based on diffusion models to detect and segment anomalies in brain imaging.
Our diffusion models achieve competitive performance compared with autoregressive approaches across a series of experiments with 2D CT and MRI data.
arXiv Detail & Related papers (2022-06-07T17:30:43Z) - Metamorphic Testing-based Adversarial Attack to Fool Deepfake Detectors [2.0649235321315285]
There is a dire need for deepfake detection technology to help spot deepfake media.
Current deepfake detection models are able to achieve outstanding accuracy (>90%)
This study identifies makeup application as an adversarial attack that could fool deepfake detectors.
arXiv Detail & Related papers (2022-04-19T02:24:30Z) - Unravelling the Effect of Image Distortions for Biased Prediction of
Pre-trained Face Recognition Models [86.79402670904338]
We evaluate the performance of four state-of-the-art deep face recognition models in the presence of image distortions.
We have observed that image distortions have a relationship with the performance gap of the model across different subgroups.
arXiv Detail & Related papers (2021-08-14T16:49:05Z) - Prediction Surface Uncertainty Quantification in Object Detection Models
for Autonomous Driving [5.784950275336468]
Object detection in autonomous cars is commonly based on camera images and Lidar inputs, which are often used to train prediction models.
We propose a novel method called PURE (Prediction sURface uncErtainty) for measuring prediction uncertainty of such regression models.
arXiv Detail & Related papers (2021-07-11T08:31:15Z) - Gradients as a Measure of Uncertainty in Neural Networks [16.80077149399317]
We propose to utilize backpropagated gradients to quantify the uncertainty of trained models.
We show that our gradient-based method outperforms state-of-the-art methods by up to 4.8% of AUROC score in out-of-distribution detection.
arXiv Detail & Related papers (2020-08-18T16:58:46Z) - How do Decisions Emerge across Layers in Neural Models? Interpretation
with Differentiable Masking [70.92463223410225]
DiffMask learns to mask-out subsets of the input while maintaining differentiability.
Decision to include or disregard an input token is made with a simple model based on intermediate hidden layers.
This lets us not only plot attribution heatmaps but also analyze how decisions are formed across network layers.
arXiv Detail & Related papers (2020-04-30T17:36:14Z) - InsideBias: Measuring Bias in Deep Networks and Application to Face
Gender Biometrics [73.85525896663371]
This work explores the biases in learning processes based on deep neural network architectures.
We employ two gender detection models based on popular deep neural networks.
We propose InsideBias, a novel method to detect biased models.
arXiv Detail & Related papers (2020-04-14T15:20:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.