They, Them, Theirs: Rewriting with Gender-Neutral English
- URL: http://arxiv.org/abs/2102.06788v1
- Date: Fri, 12 Feb 2021 21:47:48 GMT
- Title: They, Them, Theirs: Rewriting with Gender-Neutral English
- Authors: Tony Sun, Kellie Webster, Apu Shah, William Yang Wang, Melvin Johnson
- Abstract summary: We perform a case study on the singular they, a common way to promote gender inclusion in English.
We show how a model can be trained to produce gender-neutral English with 1% word error rate with no human-labeled data.
- Score: 56.14842450974887
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Responsible development of technology involves applications being inclusive
of the diverse set of users they hope to support. An important part of this is
understanding the many ways to refer to a person and being able to fluently
change between the different forms as needed. We perform a case study on the
singular they, a common way to promote gender inclusion in English. We define a
re-writing task, create an evaluation benchmark, and show how a model can be
trained to produce gender-neutral English with <1% word error rate with no
human-labeled data. We discuss the practical applications and ethical
considerations of the task, providing direction for future work into inclusive
natural language systems.
Related papers
- Beyond Binary Gender: Evaluating Gender-Inclusive Machine Translation with Ambiguous Attitude Words [85.48043537327258]
Existing machine translation gender bias evaluations are primarily focused on male and female genders.
This study presents a benchmark AmbGIMT (Gender-Inclusive Machine Translation with Ambiguous attitude words)
We propose a novel process to evaluate gender bias based on the Emotional Attitude Score (EAS), which is used to quantify ambiguous attitude words.
arXiv Detail & Related papers (2024-07-23T08:13:51Z) - Evaluation of Large Language Models: STEM education and Gender Stereotypes [2.1600200859609178]
This paper investigates gender biases in large language models (LLMs) in relation to educational choices through an open-ended, true to user-case experimental design and a quantitative analysis.
We find significant and large differences in the ratio of STEM to non-STEM suggested education paths provided by chatGPT when using typical girl vs boy names.
arXiv Detail & Related papers (2024-06-14T15:42:42Z) - Hi Guys or Hi Folks? Benchmarking Gender-Neutral Machine Translation
with the GeNTE Corpus [15.388894407006852]
Machine translation (MT) often defaults to masculine and stereotypical representations by making undue binary gender assumptions.
Our work addresses the rising demand for inclusive language by focusing head-on on gender-neutral translation from English to Italian.
arXiv Detail & Related papers (2023-10-08T21:44:00Z) - VisoGender: A dataset for benchmarking gender bias in image-text pronoun
resolution [80.57383975987676]
VisoGender is a novel dataset for benchmarking gender bias in vision-language models.
We focus on occupation-related biases within a hegemonic system of binary gender, inspired by Winograd and Winogender schemas.
We benchmark several state-of-the-art vision-language models and find that they demonstrate bias in resolving binary gender in complex scenes.
arXiv Detail & Related papers (2023-06-21T17:59:51Z) - BabySLM: language-acquisition-friendly benchmark of self-supervised
spoken language models [56.93604813379634]
Self-supervised techniques for learning speech representations have been shown to develop linguistic competence from exposure to speech without the need for human labels.
We propose a language-acquisition-friendly benchmark to probe spoken language models at the lexical and syntactic levels.
We highlight two exciting challenges that need to be addressed for further progress: bridging the gap between text and speech and between clean speech and in-the-wild speech.
arXiv Detail & Related papers (2023-06-02T12:54:38Z) - Target-Agnostic Gender-Aware Contrastive Learning for Mitigating Bias in
Multilingual Machine Translation [28.471506840241602]
Gender bias is a significant issue in machine translation, leading to ongoing research efforts in developing bias mitigation techniques.
We propose a bias mitigation method based on a novel approach.
Gender-Aware Contrastive Learning, GACL, encodes contextual gender information into the representations of non-explicit gender words.
arXiv Detail & Related papers (2023-05-23T12:53:39Z) - She Elicits Requirements and He Tests: Software Engineering Gender Bias
in Large Language Models [17.837267486473415]
This study uses data mining techniques to investigate the extent to which software development tasks are affected by implicit gender bias.
We translate each task from English into a genderless language and back, and investigate the pronouns associated with each task.
Specifically, requirements elicitation was associated with the pronoun "he" in only 6% of cases, while testing was associated with "he" in 100% of cases.
arXiv Detail & Related papers (2023-03-17T17:16:53Z) - INCLUSIFY: A benchmark and a model for gender-inclusive German [0.0]
Gender-inclusive language is important for achieving gender equality in languages with gender inflections.
A handful of tools have been developed to help people use gender-inclusive language.
We present a dataset and measures for benchmarking them, and present a model that implements these tasks.
arXiv Detail & Related papers (2022-12-05T19:37:48Z) - Analyzing Gender Representation in Multilingual Models [59.21915055702203]
We focus on the representation of gender distinctions as a practical case study.
We examine the extent to which the gender concept is encoded in shared subspaces across different languages.
arXiv Detail & Related papers (2022-04-20T00:13:01Z) - Multi-Dimensional Gender Bias Classification [67.65551687580552]
Machine learning models can inadvertently learn socially undesirable patterns when training on gender biased text.
We propose a general framework that decomposes gender bias in text along several pragmatic and semantic dimensions.
Using this fine-grained framework, we automatically annotate eight large scale datasets with gender information.
arXiv Detail & Related papers (2020-05-01T21:23:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.