Participatory Research as a Path to Community-Informed, Gender-Fair
Machine Translation
- URL: http://arxiv.org/abs/2306.08906v1
- Date: Thu, 15 Jun 2023 07:20:14 GMT
- Title: Participatory Research as a Path to Community-Informed, Gender-Fair
Machine Translation
- Authors: Dagmar Gromann, Manuel Lardelli, Katta Spiel, Sabrina Burtscher, Lukas
Daniel Klausner, Arthur Mettinger, Igor Miladinovic, Sigrid Schefer-Wenzl,
Daniela Duh, Katharina B\"uhn
- Abstract summary: We propose a method and case study building on participatory action research to include queer and non-binary people, translators, and MT experts.
The case study focuses on German, where central findings are the importance of context dependency to avoid identity invalidation.
- Score: 19.098548371499678
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent years have seen a strongly increased visibility of non-binary people
in public discourse. Accordingly, considerations of gender-fair language go
beyond a binary conception of male/female. However, language technology,
especially machine translation (MT), still suffers from binary gender bias.
Proposing a solution for gender-fair MT beyond the binary from a purely
technological perspective might fall short to accommodate different target user
groups and in the worst case might lead to misgendering. To address this
challenge, we propose a method and case study building on participatory action
research to include experiential experts, i.e., queer and non-binary people,
translators, and MT experts, in the MT design process. The case study focuses
on German, where central findings are the importance of context dependency to
avoid identity invalidation and a desire for customizable MT solutions.
Related papers
- Generating Gender Alternatives in Machine Translation [13.153018685139413]
Machine translation systems often translate terms with ambiguous gender into the gendered form that is most prevalent in the systems' training data.
This often reflects and perpetuates harmful stereotypes present in society.
We study the problem of generating all grammatically correct gendered translation alternatives.
arXiv Detail & Related papers (2024-07-29T22:10:51Z) - Beyond Binary Gender: Evaluating Gender-Inclusive Machine Translation with Ambiguous Attitude Words [85.48043537327258]
Existing machine translation gender bias evaluations are primarily focused on male and female genders.
This study presents a benchmark AmbGIMT (Gender-Inclusive Machine Translation with Ambiguous attitude words)
We propose a novel process to evaluate gender bias based on the Emotional Attitude Score (EAS), which is used to quantify ambiguous attitude words.
arXiv Detail & Related papers (2024-07-23T08:13:51Z) - Building Bridges: A Dataset for Evaluating Gender-Fair Machine Translation into German [17.924716793621627]
We study gender-fair language in English-to-German machine translation (MT)
We conduct the first benchmark study involving two commercial systems and six neural MT models.
Our findings show that most systems produce mainly masculine forms and rarely gender-neutral variants.
arXiv Detail & Related papers (2024-06-10T09:39:19Z) - Tokenization Matters: Navigating Data-Scarce Tokenization for Gender Inclusive Language Technologies [75.85462924188076]
Gender-inclusive NLP research has documented the harmful limitations of gender binary-centric large language models (LLM)
We find that misgendering is significantly influenced by Byte-Pair (BPE) tokenization.
We propose two techniques: (1) pronoun tokenization parity, a method to enforce consistent tokenization across gendered pronouns, and (2) utilizing pre-existing LLM pronoun knowledge to improve neopronoun proficiency.
arXiv Detail & Related papers (2023-12-19T01:28:46Z) - VisoGender: A dataset for benchmarking gender bias in image-text pronoun
resolution [80.57383975987676]
VisoGender is a novel dataset for benchmarking gender bias in vision-language models.
We focus on occupation-related biases within a hegemonic system of binary gender, inspired by Winograd and Winogender schemas.
We benchmark several state-of-the-art vision-language models and find that they demonstrate bias in resolving binary gender in complex scenes.
arXiv Detail & Related papers (2023-06-21T17:59:51Z) - Gender, names and other mysteries: Towards the ambiguous for
gender-inclusive translation [7.322734499960981]
This paper explores the case where the source sentence lacks explicit gender markers, but the target sentence contains them due to richer grammatical gender.
We find that many name-gender co-occurrences in MT data are not resolvable with 'unambiguous gender' in the source language.
We discuss potential steps toward gender-inclusive translation which accepts the ambiguity in both gender and translation.
arXiv Detail & Related papers (2023-06-07T16:21:59Z) - "I'm fully who I am": Towards Centering Transgender and Non-Binary
Voices to Measure Biases in Open Language Generation [69.25368160338043]
Transgender and non-binary (TGNB) individuals disproportionately experience discrimination and exclusion from daily life.
We assess how the social reality surrounding experienced marginalization of TGNB persons contributes to and persists within Open Language Generation.
We introduce TANGO, a dataset of template-based real-world text curated from a TGNB-oriented community.
arXiv Detail & Related papers (2023-05-17T04:21:45Z) - Gender Neutralization for an Inclusive Machine Translation: from
Theoretical Foundations to Open Challenges [11.37307883423629]
We explore gender-neutral translation (GNT) as a form of gender inclusivity and a goal to be achieved by machine translation (MT) models.
Specifically, we focus on translation from English into Italian, a language pair representative of salient gender-related linguistic transfer problems.
arXiv Detail & Related papers (2023-01-24T15:26:36Z) - Towards Understanding Gender-Seniority Compound Bias in Natural Language
Generation [64.65911758042914]
We investigate how seniority impacts the degree of gender bias exhibited in pretrained neural generation models.
Our results show that GPT-2 amplifies bias by considering women as junior and men as senior more often than the ground truth in both domains.
These results suggest that NLP applications built using GPT-2 may harm women in professional capacities.
arXiv Detail & Related papers (2022-05-19T20:05:02Z) - Multi-Dimensional Gender Bias Classification [67.65551687580552]
Machine learning models can inadvertently learn socially undesirable patterns when training on gender biased text.
We propose a general framework that decomposes gender bias in text along several pragmatic and semantic dimensions.
Using this fine-grained framework, we automatically annotate eight large scale datasets with gender information.
arXiv Detail & Related papers (2020-05-01T21:23:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.