Measuring Normative and Descriptive Biases in Language Models Using
Census Data
- URL: http://arxiv.org/abs/2304.05764v1
- Date: Wed, 12 Apr 2023 11:06:14 GMT
- Title: Measuring Normative and Descriptive Biases in Language Models Using
Census Data
- Authors: Samia Touileb, Lilja {\O}vrelid, Erik Velldal
- Abstract summary: We investigate how occupations with respect to gender is reflected in pre-trained language models.
We introduce an approach for measuring to what degree pre-trained language models are aligned to normative and descriptive occupational distributions.
- Score: 6.445605125467574
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We investigate in this paper how distributions of occupations with respect to
gender is reflected in pre-trained language models. Such distributions are not
always aligned to normative ideals, nor do they necessarily reflect a
descriptive assessment of reality. In this paper, we introduce an approach for
measuring to what degree pre-trained language models are aligned to normative
and descriptive occupational distributions. To this end, we use official
demographic information about gender--occupation distributions provided by the
national statistics agencies of France, Norway, United Kingdom, and the United
States. We manually generate template-based sentences combining gendered
pronouns and nouns with occupations, and subsequently probe a selection of ten
language models covering the English, French, and Norwegian languages. The
scoring system we introduce in this work is language independent, and can be
used on any combination of template-based sentences, occupations, and
languages. The approach could also be extended to other dimensions of national
census data and other demographic variables.
Related papers
- Forcing Diffuse Distributions out of Language Models [70.28345569190388]
Despite being trained specifically to follow user instructions, today's instructiontuned language models perform poorly when instructed to produce random outputs.
We propose a fine-tuning method that encourages language models to output distributions that are diffuse over valid outcomes.
arXiv Detail & Related papers (2024-04-16T19:17:23Z) - What is Your Favorite Gender, MLM? Gender Bias Evaluation in Multilingual Masked Language Models [8.618945530676614]
This paper proposes an approach to estimate gender bias in multilingual lexicons from 5 languages: Chinese, English, German, Portuguese, and Spanish.
A novel model-based method is presented to generate sentence pairs for a more robust analysis of gender bias.
Our results suggest that gender bias should be studied on a large dataset using multiple evaluation metrics for best practice.
arXiv Detail & Related papers (2024-04-09T21:12:08Z) - Multilingual Text-to-Image Generation Magnifies Gender Stereotypes and Prompt Engineering May Not Help You [64.74707085021858]
We show that multilingual models suffer from significant gender biases just as monolingual models do.
We propose a novel benchmark, MAGBIG, intended to foster research on gender bias in multilingual models.
Our results show that not only do models exhibit strong gender biases but they also behave differently across languages.
arXiv Detail & Related papers (2024-01-29T12:02:28Z) - Cross-Lingual NER for Financial Transaction Data in Low-Resource
Languages [70.25418443146435]
We propose an efficient modeling framework for cross-lingual named entity recognition in semi-structured text data.
We employ two independent datasets of SMSs in English and Arabic, each carrying semi-structured banking transaction information.
With access to only 30 labeled samples, our model can generalize the recognition of merchants, amounts, and other fields from English to Arabic.
arXiv Detail & Related papers (2023-07-16T00:45:42Z) - Multilingual Conceptual Coverage in Text-to-Image Models [98.80343331645626]
"Conceptual Coverage Across Languages" (CoCo-CroLa) is a technique for benchmarking the degree to which any generative text-to-image system provides multilingual parity to its training language in terms of tangible nouns.
For each model we can assess "conceptual coverage" of a given target language relative to a source language by comparing the population of images generated for a series of tangible nouns in the source language to the population of images generated for each noun under translation in the target language.
arXiv Detail & Related papers (2023-06-02T17:59:09Z) - INCLUSIFY: A benchmark and a model for gender-inclusive German [0.0]
Gender-inclusive language is important for achieving gender equality in languages with gender inflections.
A handful of tools have been developed to help people use gender-inclusive language.
We present a dataset and measures for benchmarking them, and present a model that implements these tasks.
arXiv Detail & Related papers (2022-12-05T19:37:48Z) - An Analysis of Social Biases Present in BERT Variants Across Multiple
Languages [0.0]
We investigate the bias present in monolingual BERT models across a diverse set of languages.
We propose a template-based method to measure any kind of bias, based on sentence pseudo-likelihood.
We conclude that current methods of probing for bias are highly language-dependent.
arXiv Detail & Related papers (2022-11-25T23:38:08Z) - Efficient Gender Debiasing of Pre-trained Indic Language Models [0.0]
The gender bias present in the data on which language models are pre-trained gets reflected in the systems that use these models.
In our paper, we measure gender bias associated with occupations in Hindi language models.
Our results reflect that the bias is reduced post-introduction of our proposed mitigation techniques.
arXiv Detail & Related papers (2022-09-08T09:15:58Z) - Evaluating Gender Bias in Hindi-English Machine Translation [0.1503974529275767]
We implement a modified version of the TGBI metric based on the grammatical considerations for Hindi.
We compare and contrast the resulting bias measurements across multiple metrics for pre-trained embeddings and the ones learned by our machine translation model.
arXiv Detail & Related papers (2021-06-16T10:35:51Z) - They, Them, Theirs: Rewriting with Gender-Neutral English [56.14842450974887]
We perform a case study on the singular they, a common way to promote gender inclusion in English.
We show how a model can be trained to produce gender-neutral English with 1% word error rate with no human-labeled data.
arXiv Detail & Related papers (2021-02-12T21:47:48Z) - Multi-Dimensional Gender Bias Classification [67.65551687580552]
Machine learning models can inadvertently learn socially undesirable patterns when training on gender biased text.
We propose a general framework that decomposes gender bias in text along several pragmatic and semantic dimensions.
Using this fine-grained framework, we automatically annotate eight large scale datasets with gender information.
arXiv Detail & Related papers (2020-05-01T21:23:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.