Quantifying Stereotypes in Language
- URL: http://arxiv.org/abs/2401.15535v1
- Date: Sun, 28 Jan 2024 01:07:21 GMT
- Title: Quantifying Stereotypes in Language
- Authors: Yang Liu
- Abstract summary: We quantify stereotypes in language by annotating a dataset.
We use the pre-trained language models (PLMs) to learn this dataset to predict stereotypes of sentences.
We discuss stereotypes about common social issues such as hate speech, sexism, sentiments, and disadvantaged and advantaged groups.
- Score: 6.697298321551588
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A stereotype is a generalized perception of a specific group of humans. It is
often potentially encoded in human language, which is more common in texts on
social issues. Previous works simply define a sentence as stereotypical and
anti-stereotypical. However, the stereotype of a sentence may require
fine-grained quantification. In this paper, to fill this gap, we quantify
stereotypes in language by annotating a dataset. We use the pre-trained
language models (PLMs) to learn this dataset to predict stereotypes of
sentences. Then, we discuss stereotypes about common social issues such as hate
speech, sexism, sentiments, and disadvantaged and advantaged groups. We
demonstrate the connections and differences between stereotypes and common
social issues, and all four studies validate the general findings of the
current studies. In addition, our work suggests that fine-grained stereotype
scores are a highly relevant and competitive dimension for research on social
issues.
Related papers
- Spoken Stereoset: On Evaluating Social Bias Toward Speaker in Speech Large Language Models [50.40276881893513]
This study introduces Spoken Stereoset, a dataset specifically designed to evaluate social biases in Speech Large Language Models (SLLMs)
By examining how different models respond to speech from diverse demographic groups, we aim to identify these biases.
The findings indicate that while most models show minimal bias, some still exhibit slightly stereotypical or anti-stereotypical tendencies.
arXiv Detail & Related papers (2024-08-14T16:55:06Z) - GeniL: A Multilingual Dataset on Generalizing Language [19.43611224855484]
Current methods to assess presence of stereotypes in generated language rely on simple template or co-occurrence based measures.
We argue that understanding the sentential context is crucial for detecting instances of generalization.
We build GeniL, a multilingual dataset of over 50K sentences from 9 languages annotated for instances of generalizations.
arXiv Detail & Related papers (2024-04-08T20:58:06Z) - Comparing Biases and the Impact of Multilingual Training across Multiple
Languages [70.84047257764405]
We present a bias analysis across Italian, Chinese, English, Hebrew, and Spanish on the downstream sentiment analysis task.
We adapt existing sentiment bias templates in English to Italian, Chinese, Hebrew, and Spanish for four attributes: race, religion, nationality, and gender.
Our results reveal similarities in bias expression such as favoritism of groups that are dominant in each language's culture.
arXiv Detail & Related papers (2023-05-18T18:15:07Z) - Easily Accessible Text-to-Image Generation Amplifies Demographic
Stereotypes at Large Scale [61.555788332182395]
We investigate the potential for machine learning models to amplify dangerous and complex stereotypes.
We find a broad range of ordinary prompts produce stereotypes, including prompts simply mentioning traits, descriptors, occupations, or objects.
arXiv Detail & Related papers (2022-11-07T18:31:07Z) - Hate Speech Classifiers Learn Human-Like Social Stereotypes [4.132204773132937]
Social stereotypes negatively impact individuals' judgements about different groups.
Social stereotypes may have a critical role in how people understand language directed toward minority social groups.
arXiv Detail & Related papers (2021-10-28T01:35:41Z) - Stepmothers are mean and academics are pretentious: What do pretrained
language models learn about you? [11.107926166222452]
We present the first dataset comprising stereotypical attributes of a range of social groups.
We propose a method to elicit stereotypes encoded by pretrained language models in an unsupervised fashion.
arXiv Detail & Related papers (2021-09-21T09:44:57Z) - Towards Understanding and Mitigating Social Biases in Language Models [107.82654101403264]
Large-scale pretrained language models (LMs) can be potentially dangerous in manifesting undesirable representational biases.
We propose steps towards mitigating social biases during text generation.
Our empirical results and human evaluation demonstrate effectiveness in mitigating bias while retaining crucial contextual information.
arXiv Detail & Related papers (2021-06-24T17:52:43Z) - Understanding and Countering Stereotypes: A Computational Approach to
the Stereotype Content Model [4.916009028580767]
We present a computational approach to interpreting stereotypes in text through the Stereotype Content Model (SCM)
The SCM proposes that stereotypes can be understood along two primary dimensions: warmth and competence.
It is known that countering stereotypes with anti-stereotypical examples is one of the most effective ways to reduce biased thinking.
arXiv Detail & Related papers (2021-06-04T16:53:37Z) - UnQovering Stereotyping Biases via Underspecified Questions [68.81749777034409]
We present UNQOVER, a framework to probe and quantify biases through underspecified questions.
We show that a naive use of model scores can lead to incorrect bias estimates due to two forms of reasoning errors.
We use this metric to analyze four important classes of stereotypes: gender, nationality, ethnicity, and religion.
arXiv Detail & Related papers (2020-10-06T01:49:52Z) - CrowS-Pairs: A Challenge Dataset for Measuring Social Biases in Masked
Language Models [30.582132471411263]
We introduce the Crowd Stereotype Pairs benchmark (CrowS-Pairs)
CrowS-Pairs has 1508 examples that cover stereotypes dealing with nine types of bias, like race, religion, and age.
We find that all three of the widely-used sentences we evaluate substantially favor stereotypes in every category in CrowS-Pairs.
arXiv Detail & Related papers (2020-09-30T22:38:40Z) - Towards Controllable Biases in Language Generation [87.89632038677912]
We develop a method to induce societal biases in generated text when input prompts contain mentions of specific demographic groups.
We analyze two scenarios: 1) inducing negative biases for one demographic and positive biases for another demographic, and 2) equalizing biases between demographics.
arXiv Detail & Related papers (2020-05-01T08:25:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.