Gender Stereotype Reinforcement: Measuring the Gender Bias Conveyed by
Ranking Algorithms
- URL: http://arxiv.org/abs/2009.01334v1
- Date: Wed, 2 Sep 2020 20:45:04 GMT
- Title: Gender Stereotype Reinforcement: Measuring the Gender Bias Conveyed by
Ranking Algorithms
- Authors: Alessandro Fabris, Alberto Purpura, Gianmaria Silvello, Gian Antonio
Susto
- Abstract summary: We propose the Gender Stereotype Reinforcement (GSR) measure, which quantifies the tendency of a Search Engines to support gender stereotypes.
GSR is the first specifically tailored measure for Information Retrieval, capable of quantifying representational harms.
- Score: 68.85295025020942
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Search Engines (SE) have been shown to perpetuate well-known gender
stereotypes identified in psychology literature and to influence users
accordingly. Similar biases were found encoded in Word Embeddings (WEs) learned
from large online corpora. In this context, we propose the Gender Stereotype
Reinforcement (GSR) measure, which quantifies the tendency of a SE to support
gender stereotypes, leveraging gender-related information encoded in WEs.
Through the critical lens of construct validity, we validate the proposed
measure on synthetic and real collections. Subsequently, we use GSR to compare
widely-used Information Retrieval ranking algorithms, including lexical,
semantic, and neural models. We check if and how ranking algorithms based on
WEs inherit the biases of the underlying embeddings. We also consider the most
common debiasing approaches for WEs proposed in the literature and test their
impact in terms of GSR and common performance measures. To the best of our
knowledge, GSR is the first specifically tailored measure for IR, capable of
quantifying representational harms.
Related papers
- The Root Shapes the Fruit: On the Persistence of Gender-Exclusive Harms in Aligned Language Models [58.130894823145205]
We center transgender, nonbinary, and other gender-diverse identities to investigate how alignment procedures interact with pre-existing gender-diverse bias.
Our findings reveal that DPO-aligned models are particularly sensitive to supervised finetuning.
We conclude with recommendations tailored to DPO and broader alignment practices.
arXiv Detail & Related papers (2024-11-06T06:50:50Z) - GenderCARE: A Comprehensive Framework for Assessing and Reducing Gender Bias in Large Language Models [73.23743278545321]
Large language models (LLMs) have exhibited remarkable capabilities in natural language generation, but have also been observed to magnify societal biases.
GenderCARE is a comprehensive framework that encompasses innovative Criteria, bias Assessment, Reduction techniques, and Evaluation metrics.
arXiv Detail & Related papers (2024-08-22T15:35:46Z) - Towards Region-aware Bias Evaluation Metrics [26.91545185271231]
We identify topical differences in gender bias across different regions and propose a region-aware bottom-up approach for bias assessment.
Our proposed approach uses gender-aligned topics for a given region and identifies gender bias dimensions in the form of topic pairs.
Several of our proposed bias topic pairs are on par with human perception of gender biases in these regions in comparison to the existing ones.
arXiv Detail & Related papers (2024-06-23T16:26:27Z) - ''Fifty Shades of Bias'': Normative Ratings of Gender Bias in GPT
Generated English Text [11.085070600065801]
Language serves as a powerful tool for the manifestation of societal belief systems.
Gender bias is one of the most pervasive biases in our society.
We create the first dataset of GPT-generated English text with normative ratings of gender bias.
arXiv Detail & Related papers (2023-10-26T14:34:06Z) - Diversity matters: Robustness of bias measurements in Wikidata [4.950095974653716]
We reveal data biases that surface in Wikidata for thirteen different demographics selected from seven continents.
We conduct our extensive experiments on a large number of occupations sampled from the thirteen demographics with respect to the sensitive attribute, i.e., gender.
We show that the choice of the state-of-the-art KG embedding algorithm has a strong impact on the ranking of biased occupations irrespective of gender.
arXiv Detail & Related papers (2023-02-27T18:38:10Z) - Auditing Gender Presentation Differences in Text-to-Image Models [54.16959473093973]
We study how gender is presented differently in text-to-image models.
By probing gender indicators in the input text, we quantify the frequency differences of presentation-centric attributes.
We propose an automatic method to estimate such differences.
arXiv Detail & Related papers (2023-02-07T18:52:22Z) - Social Biases in Automatic Evaluation Metrics for NLG [53.76118154594404]
We propose an evaluation method based on Word Embeddings Association Test (WEAT) and Sentence Embeddings Association Test (SEAT) to quantify social biases in evaluation metrics.
We construct gender-swapped meta-evaluation datasets to explore the potential impact of gender bias in image caption and text summarization tasks.
arXiv Detail & Related papers (2022-10-17T08:55:26Z) - Gender Bias in Text: Labeled Datasets and Lexicons [0.30458514384586394]
There is a lack of gender bias datasets and lexicons for automating the detection of gender bias.
We provide labeled datasets and exhaustive lexicons by collecting, annotating, and augmenting relevant sentences.
The released datasets and lexicons span multiple bias subtypes including: Generic He, Generic She, Explicit Marking of Sex, and Gendered Neologisms.
arXiv Detail & Related papers (2022-01-21T12:44:51Z) - Social Norm Bias: Residual Harms of Fairness-Aware Algorithms [21.50551404445654]
Social Norm Bias (SNoB) is a subtle but consequential type of discrimination that may be exhibited by automated decision-making systems.
We quantify SNoB by measuring how an algorithm's predictions are associated with conformity to gender norms.
We show that post-processing interventions do not mitigate this type of bias at all.
arXiv Detail & Related papers (2021-08-25T05:54:56Z) - Multi-Dimensional Gender Bias Classification [67.65551687580552]
Machine learning models can inadvertently learn socially undesirable patterns when training on gender biased text.
We propose a general framework that decomposes gender bias in text along several pragmatic and semantic dimensions.
Using this fine-grained framework, we automatically annotate eight large scale datasets with gender information.
arXiv Detail & Related papers (2020-05-01T21:23:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.