Are Commercial Face Detection Models as Biased as Academic Models?
- URL: http://arxiv.org/abs/2201.10047v2
- Date: Wed, 30 Nov 2022 02:14:23 GMT
- Title: Are Commercial Face Detection Models as Biased as Academic Models?
- Authors: Samuel Dooley, George Z. Wei, Tom Goldstein, John P. Dickerson
- Abstract summary: We compare academic and commercial face detection systems, specifically examining robustness to noise.
We find that state-of-the-art academic face detection models exhibit demographic disparities in their noise robustness.
We conclude that commercial models are always as biased or more biased than an academic model.
- Score: 64.71318433419636
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As facial recognition systems are deployed more widely, scholars and
activists have studied their biases and harms. Audits are commonly used to
accomplish this and compare the algorithmic facial recognition systems'
performance against datasets with various metadata labels about the subjects of
the images. Seminal works have found discrepancies in performance by gender
expression, age, perceived race, skin type, etc. These studies and audits often
examine algorithms which fall into two categories: academic models or
commercial models. We present a detailed comparison between academic and
commercial face detection systems, specifically examining robustness to noise.
We find that state-of-the-art academic face detection models exhibit
demographic disparities in their noise robustness, specifically by having
statistically significant decreased performance on older individuals and those
who present their gender in a masculine manner. When we compare the size of
these disparities to that of commercial models, we conclude that commercial
models - in contrast to their relatively larger development budget and
industry-level fairness commitments - are always as biased or more biased than
an academic model.
Related papers
- Classes Are Not Equal: An Empirical Study on Image Recognition Fairness [100.36114135663836]
We experimentally demonstrate that classes are not equal and the fairness issue is prevalent for image classification models across various datasets.
Our findings reveal that models tend to exhibit greater prediction biases for classes that are more challenging to recognize.
Data augmentation and representation learning algorithms improve overall performance by promoting fairness to some degree in image classification.
arXiv Detail & Related papers (2024-02-28T07:54:50Z) - Bias and Diversity in Synthetic-based Face Recognition [12.408456748469426]
We investigate how the diversity of synthetic face recognition datasets compares to authentic datasets.
We look at the distribution of gender, ethnicity, age, and head position.
With regard to bias, it can be seen that the synthetic-based models share a similar bias behavior with the authentic-based models.
arXiv Detail & Related papers (2023-11-07T13:12:34Z) - FACET: Fairness in Computer Vision Evaluation Benchmark [21.862644380063756]
Computer vision models have known performance disparities across attributes such as gender and skin tone.
We present a new benchmark named FACET (FAirness in Computer Vision EvaluaTion)
FACET is a large, publicly available evaluation set of 32k images for some of the most common vision tasks.
arXiv Detail & Related papers (2023-08-31T17:59:48Z) - Gender Stereotyping Impact in Facial Expression Recognition [1.5340540198612824]
In recent years, machine learning-based models have become the most popular approach to Facial Expression Recognition (FER)
In publicly available FER datasets, apparent gender representation is usually mostly balanced, but their representation in the individual label is not.
We generate derivative datasets with different amounts of stereotypical bias by altering the gender proportions of certain labels.
We observe a discrepancy in the recognition of certain emotions between genders of up to $29 %$ under the worst bias conditions.
arXiv Detail & Related papers (2022-10-11T10:52:23Z) - Explaining Bias in Deep Face Recognition via Image Characteristics [9.569575076277523]
We evaluate ten state-of-the-art face recognition models, comparing their fairness in terms of security and usability on two data sets.
We then analyze the impact of image characteristics on models performance.
arXiv Detail & Related papers (2022-08-23T17:18:23Z) - Comparing Human and Machine Bias in Face Recognition [46.170389064229354]
We release improvements to the LFW and CelebA datasets which will enable future researchers to obtain measurements of algorithmic bias.
We also use these new data to develop a series of challenging facial identification and verification questions.
We find that both computer models and human survey participants perform significantly better at the verification task.
arXiv Detail & Related papers (2021-10-15T22:26:20Z) - Robustness Disparities in Commercial Face Detection [72.25318723264215]
We present the first of its kind detailed benchmark of the robustness of three such systems: Amazon Rekognition, Microsoft Azure, and Google Cloud Platform.
We generally find that photos of individuals who are older, masculine presenting, of darker skin type, or have dim lighting are more susceptible to errors than their counterparts in other identities.
arXiv Detail & Related papers (2021-08-27T21:37:16Z) - Unravelling the Effect of Image Distortions for Biased Prediction of
Pre-trained Face Recognition Models [86.79402670904338]
We evaluate the performance of four state-of-the-art deep face recognition models in the presence of image distortions.
We have observed that image distortions have a relationship with the performance gap of the model across different subgroups.
arXiv Detail & Related papers (2021-08-14T16:49:05Z) - Investigating Bias in Deep Face Analysis: The KANFace Dataset and
Empirical Study [67.3961439193994]
We introduce the most comprehensive, large-scale dataset of facial images and videos to date.
The data are manually annotated in terms of identity, exact age, gender and kinship.
A method to debias network embeddings is introduced and tested on the proposed benchmarks.
arXiv Detail & Related papers (2020-05-15T00:14:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.