Markedness in Visual Semantic AI
- URL: http://arxiv.org/abs/2205.11378v1
- Date: Mon, 23 May 2022 15:14:41 GMT
- Title: Markedness in Visual Semantic AI
- Authors: Robert Wolfe, Aylin Caliskan
- Abstract summary: We evaluate the state-of-the-art multimodal "visual semantic" model CLIP for biases related to the marking of age, gender, and race or ethnicity.
Female individuals under the age of 20 are more likely than Male individuals to be marked with a gender label, but less likely to be marked with an age label.
As age increases, the self-similarity of representations of Female individuals increases at a higher rate than for Male individuals.
- Score: 3.4157048274143316
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: We evaluate the state-of-the-art multimodal "visual semantic" model CLIP
("Contrastive Language Image Pretraining") for biases related to the marking of
age, gender, and race or ethnicity. Given the option to label an image as "a
photo of a person" or to select a label denoting race or ethnicity, CLIP
chooses the "person" label 47.9% of the time for White individuals, compared
with 5.0% or less for individuals who are Black, East Asian, Southeast Asian,
Indian, or Latino or Hispanic. The model is more likely to rank the unmarked
"person" label higher than labels denoting gender for Male individuals (26.7%
of the time) vs. Female individuals (15.2% of the time). Age affects whether an
individual is marked by the model: Female individuals under the age of 20 are
more likely than Male individuals to be marked with a gender label, but less
likely to be marked with an age label, while Female individuals over the age of
40 are more likely to be marked based on age than Male individuals. We also
examine the self-similarity (mean pairwise cosine similarity) for each social
group, where higher self-similarity denotes greater attention directed by CLIP
to the shared characteristics (age, race, or gender) of the social group. As
age increases, the self-similarity of representations of Female individuals
increases at a higher rate than for Male individuals, with the disparity most
pronounced at the "more than 70" age range. All ten of the most self-similar
social groups are individuals under the age of 10 or over the age of 70, and
six of the ten are Female individuals. Existing biases of self-similarity and
markedness between Male and Female gender groups are further exacerbated when
the groups compared are individuals who are White and Male and individuals who
are Black and Female. Results indicate that CLIP reflects the biases of the
language and society which produced its training data.
Related papers
- GenderBias-\emph{VL}: Benchmarking Gender Bias in Vision Language Models via Counterfactual Probing [72.0343083866144]
This paper introduces the GenderBias-emphVL benchmark to evaluate occupation-related gender bias in Large Vision-Language Models.
Using our benchmark, we extensively evaluate 15 commonly used open-source LVLMs and state-of-the-art commercial APIs.
Our findings reveal widespread gender biases in existing LVLMs.
arXiv Detail & Related papers (2024-06-30T05:55:15Z) - AI-generated faces influence gender stereotypes and racial homogenization [1.6647208383676708]
We document significant biases in Stable Diffusion across six races, two genders, 32 professions, and eight attributes.
This analysis reveals significant racial homogenization depicting nearly all Middle Eastern men as bearded, brown-skinned, and wearing traditional attire.
We propose debiasing solutions that allow users to specify the desired distributions of race and gender when generating images.
arXiv Detail & Related papers (2024-02-01T20:32:14Z) - Large Language Models Portray Socially Subordinate Groups as More Homogeneous, Consistent with a Bias Observed in Humans [0.30723404270319693]
We investigate a new form of bias in large language models (LLMs)
We find that ChatGPT portrayed African, Asian, and Hispanic Americans as more homogeneous than White Americans.
We argue that the tendency to describe groups as less diverse risks perpetuating stereotypes and discriminatory behavior.
arXiv Detail & Related papers (2024-01-16T16:52:00Z) - VisoGender: A dataset for benchmarking gender bias in image-text pronoun
resolution [80.57383975987676]
VisoGender is a novel dataset for benchmarking gender bias in vision-language models.
We focus on occupation-related biases within a hegemonic system of binary gender, inspired by Winograd and Winogender schemas.
We benchmark several state-of-the-art vision-language models and find that they demonstrate bias in resolving binary gender in complex scenes.
arXiv Detail & Related papers (2023-06-21T17:59:51Z) - Gender Stereotyping Impact in Facial Expression Recognition [1.5340540198612824]
In recent years, machine learning-based models have become the most popular approach to Facial Expression Recognition (FER)
In publicly available FER datasets, apparent gender representation is usually mostly balanced, but their representation in the individual label is not.
We generate derivative datasets with different amounts of stereotypical bias by altering the gender proportions of certain labels.
We observe a discrepancy in the recognition of certain emotions between genders of up to $29 %$ under the worst bias conditions.
arXiv Detail & Related papers (2022-10-11T10:52:23Z) - Are Commercial Face Detection Models as Biased as Academic Models? [64.71318433419636]
We compare academic and commercial face detection systems, specifically examining robustness to noise.
We find that state-of-the-art academic face detection models exhibit demographic disparities in their noise robustness.
We conclude that commercial models are always as biased or more biased than an academic model.
arXiv Detail & Related papers (2022-01-25T02:21:42Z) - Fairness for Image Generation with Uncertain Sensitive Attributes [97.81354305427871]
This work tackles the issue of fairness in the context of generative procedures, such as image super-resolution.
While traditional group fairness definitions are typically defined with respect to specified protected groups, we emphasize that there are no ground truth identities.
We show that the natural extension of demographic parity is strongly dependent on the grouping, and emphimpossible to achieve obliviously.
arXiv Detail & Related papers (2021-06-23T06:17:17Z) - Quantifying the Extent to Which Race and Gender Features Determine
Identity in Commercial Face Recognition Algorithms [0.0]
Black-box commercial face recognition algorithms (CFRAs) use gender and race features to determine identity.
This study quantified the degree to which gender and race features influenced face recognition similarity scores between different people.
arXiv Detail & Related papers (2020-10-15T18:52:36Z) - Understanding Fairness of Gender Classification Algorithms Across
Gender-Race Groups [0.8594140167290097]
The aim of this paper is to investigate the differential performance of the gender classification algorithms across gender-race groups.
For all the algorithms used, Black females (Black race in general) always obtained the least accuracy rates.
Middle Eastern males and Latino females obtained higher accuracy rates most of the time.
arXiv Detail & Related papers (2020-09-24T04:56:10Z) - Enhancing Facial Data Diversity with Style-based Face Aging [59.984134070735934]
In particular, face datasets are typically biased in terms of attributes such as gender, age, and race.
We propose a novel, generative style-based architecture for data augmentation that captures fine-grained aging patterns.
We show that the proposed method outperforms state-of-the-art algorithms for age transfer.
arXiv Detail & Related papers (2020-06-06T21:53:44Z) - Multi-Dimensional Gender Bias Classification [67.65551687580552]
Machine learning models can inadvertently learn socially undesirable patterns when training on gender biased text.
We propose a general framework that decomposes gender bias in text along several pragmatic and semantic dimensions.
Using this fine-grained framework, we automatically annotate eight large scale datasets with gender information.
arXiv Detail & Related papers (2020-05-01T21:23:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.