Auditing Gender Analyzers on Text Data
- URL: http://arxiv.org/abs/2310.06061v1
- Date: Mon, 9 Oct 2023 18:13:07 GMT
- Title: Auditing Gender Analyzers on Text Data
- Authors: Siddharth D Jaiswal, Ankit Kumar Verma, Animesh Mukherjee
- Abstract summary: We audit three existing gender analyzers -- uClassify, Readable and HackerFactor, for biases against non-binary individuals.
The tools are designed to predict only the cisgender binary labels, which leads to discrimination against non-binary members of the society.
To address this, we fine-tune a BERT multi-label classifier on the two datasets in multiple combinations.
- Score: 7.73812434373948
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: AI models have become extremely popular and accessible to the general public.
However, they are continuously under the scanner due to their demonstrable
biases toward various sections of the society like people of color and
non-binary people. In this study, we audit three existing gender analyzers --
uClassify, Readable and HackerFactor, for biases against non-binary
individuals. These tools are designed to predict only the cisgender binary
labels, which leads to discrimination against non-binary members of the
society. We curate two datasets -- Reddit comments (660k) and, Tumblr posts
(2.05M) and our experimental evaluation shows that the tools are highly
inaccurate with the overall accuracy being ~50% on all platforms. Predictions
for non-binary comments on all platforms are mostly female, thus propagating
the societal bias that non-binary individuals are effeminate. To address this,
we fine-tune a BERT multi-label classifier on the two datasets in multiple
combinations, observe an overall performance of ~77% on the most realistically
deployable setting and a surprisingly higher performance of 90% for the
non-binary class. We also audit ChatGPT using zero-shot prompts on a small
dataset (due to high pricing) and observe an average accuracy of 58% for Reddit
and Tumblr combined (with overall better results for Reddit).
Thus, we show that existing systems, including highly advanced ones like
ChatGPT are biased, and need better audits and moderation and, that such
societal biases can be addressed and alleviated through simple off-the-shelf
models like BERT trained on more gender inclusive datasets.
Related papers
- Revealing and Reducing Gender Biases in Vision and Language Assistants (VLAs) [82.57490175399693]
We study gender bias in 22 popular image-to-text vision-language assistants (VLAs)
Our results show that VLAs replicate human biases likely present in the data, such as real-world occupational imbalances.
To eliminate the gender bias in these models, we find that fine-tuning-based debiasing methods achieve the best trade-off between debiasing and retaining performance.
arXiv Detail & Related papers (2024-10-25T05:59:44Z) - Beyond Binary Gender: Evaluating Gender-Inclusive Machine Translation with Ambiguous Attitude Words [85.48043537327258]
Existing machine translation gender bias evaluations are primarily focused on male and female genders.
This study presents a benchmark AmbGIMT (Gender-Inclusive Machine Translation with Ambiguous attitude words)
We propose a novel process to evaluate gender bias based on the Emotional Attitude Score (EAS), which is used to quantify ambiguous attitude words.
arXiv Detail & Related papers (2024-07-23T08:13:51Z) - The Male CEO and the Female Assistant: Evaluation and Mitigation of Gender Biases in Text-To-Image Generation of Dual Subjects [58.27353205269664]
We propose the Paired Stereotype Test (PST) framework, which queries T2I models to depict two individuals assigned with male-stereotyped and female-stereotyped social identities.
PST queries T2I models to depict two individuals assigned with male-stereotyped and female-stereotyped social identities.
Using PST, we evaluate two aspects of gender biases -- the well-known bias in gendered occupation and a novel aspect: bias in organizational power.
arXiv Detail & Related papers (2024-02-16T21:32:27Z) - ''Fifty Shades of Bias'': Normative Ratings of Gender Bias in GPT
Generated English Text [11.085070600065801]
Language serves as a powerful tool for the manifestation of societal belief systems.
Gender bias is one of the most pervasive biases in our society.
We create the first dataset of GPT-generated English text with normative ratings of gender bias.
arXiv Detail & Related papers (2023-10-26T14:34:06Z) - "I'm fully who I am": Towards Centering Transgender and Non-Binary
Voices to Measure Biases in Open Language Generation [69.25368160338043]
Transgender and non-binary (TGNB) individuals disproportionately experience discrimination and exclusion from daily life.
We assess how the social reality surrounding experienced marginalization of TGNB persons contributes to and persists within Open Language Generation.
We introduce TANGO, a dataset of template-based real-world text curated from a TGNB-oriented community.
arXiv Detail & Related papers (2023-05-17T04:21:45Z) - Exploring Gender Bias in Retrieval Models [2.594412743115663]
Mitigating gender bias in information retrieval is important to avoid propagating stereotypes.
We employ a dataset consisting of two components: (1) relevance of a document to a query and (2) "gender" of a document.
We show that pre-trained models for IR do not perform well in zero-shot retrieval tasks when full fine-tuning of a large pre-trained BERT encoder is performed.
We also illustrate that pre-trained models have gender biases that result in retrieved articles tending to be more often male than female.
arXiv Detail & Related papers (2022-08-02T21:12:05Z) - Towards Understanding Gender-Seniority Compound Bias in Natural Language
Generation [64.65911758042914]
We investigate how seniority impacts the degree of gender bias exhibited in pretrained neural generation models.
Our results show that GPT-2 amplifies bias by considering women as junior and men as senior more often than the ground truth in both domains.
These results suggest that NLP applications built using GPT-2 may harm women in professional capacities.
arXiv Detail & Related papers (2022-05-19T20:05:02Z) - Stereotype and Skew: Quantifying Gender Bias in Pre-trained and
Fine-tuned Language Models [5.378664454650768]
This paper proposes two intuitive metrics, skew and stereotype, that quantify and analyse the gender bias present in contextual language models.
We find evidence that gender stereotype correlates approximately negatively with gender skew in out-of-the-box models, suggesting that there is a trade-off between these two forms of bias.
arXiv Detail & Related papers (2021-01-24T10:57:59Z) - Transgender Community Sentiment Analysis from Social Media Data: A
Natural Language Processing Approach [3.044968666863866]
Transgender community is experiencing a huge disparity in mental health conditions compared with the general population.
In this study, we manually categorize 300 social media comments posted by transgender people to the sentiment of negative, positive, and neutral.
arXiv Detail & Related papers (2020-10-25T08:13:34Z) - Gender Classification and Bias Mitigation in Facial Images [7.438105108643341]
Recent research showed that algorithms trained on biased benchmark databases could result in algorithmic bias.
We conducted surveys on existing benchmark databases for facial recognition and gender classification tasks.
We worked to increase classification accuracy and mitigate algorithmic biases on our baseline model trained on the augmented benchmark database.
arXiv Detail & Related papers (2020-07-13T01:09:06Z) - Mitigating Gender Bias in Captioning Systems [56.25457065032423]
Most captioning models learn gender bias, leading to high gender prediction errors, especially for women.
We propose a new Guided Attention Image Captioning model (GAIC) which provides self-guidance on visual attention to encourage the model to capture correct gender visual evidence.
arXiv Detail & Related papers (2020-06-15T12:16:19Z) - Multi-Dimensional Gender Bias Classification [67.65551687580552]
Machine learning models can inadvertently learn socially undesirable patterns when training on gender biased text.
We propose a general framework that decomposes gender bias in text along several pragmatic and semantic dimensions.
Using this fine-grained framework, we automatically annotate eight large scale datasets with gender information.
arXiv Detail & Related papers (2020-05-01T21:23:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.