Social Biases in NLP Models as Barriers for Persons with Disabilities
- URL: http://arxiv.org/abs/2005.00813v1
- Date: Sat, 2 May 2020 12:16:54 GMT
- Title: Social Biases in NLP Models as Barriers for Persons with Disabilities
- Authors: Ben Hutchinson, Vinodkumar Prabhakaran, Emily Denton, Kellie Webster,
Yu Zhong, Stephen Denuyl
- Abstract summary: We present evidence of undesirable biases towards mentions of disability in two different English language models: toxicity prediction and sentiment analysis.
Next, we demonstrate that the neural embeddings that are the critical first step in most NLP pipelines similarly contain undesirable biases towards mentions of disability.
We end by highlighting topical biases in the discourse about disability which may contribute to the observed model biases.
- Score: 13.579848462349192
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Building equitable and inclusive NLP technologies demands consideration of
whether and how social attitudes are represented in ML models. In particular,
representations encoded in models often inadvertently perpetuate undesirable
social biases from the data on which they are trained. In this paper, we
present evidence of such undesirable biases towards mentions of disability in
two different English language models: toxicity prediction and sentiment
analysis. Next, we demonstrate that the neural embeddings that are the critical
first step in most NLP pipelines similarly contain undesirable biases towards
mentions of disability. We end by highlighting topical biases in the discourse
about disability which may contribute to the observed model biases; for
instance, gun violence, homelessness, and drug addiction are over-represented
in texts discussing mental illness.
Related papers
- The Devil is in the Neurons: Interpreting and Mitigating Social Biases in Pre-trained Language Models [78.69526166193236]
Pre-trained Language models (PLMs) have been acknowledged to contain harmful information, such as social biases.
We propose sc Social Bias Neurons to accurately pinpoint units (i.e., neurons) in a language model that can be attributed to undesirable behavior, such as social bias.
As measured by prior metrics from StereoSet, our model achieves a higher degree of fairness while maintaining language modeling ability with low cost.
arXiv Detail & Related papers (2024-06-14T15:41:06Z) - Decoding Susceptibility: Modeling Misbelief to Misinformation Through a Computational Approach [61.04606493712002]
Susceptibility to misinformation describes the degree of belief in unverifiable claims that is not observable.
Existing susceptibility studies heavily rely on self-reported beliefs.
We propose a computational approach to model users' latent susceptibility levels.
arXiv Detail & Related papers (2023-11-16T07:22:56Z) - Social Bias Probing: Fairness Benchmarking for Language Models [38.180696489079985]
This paper proposes a novel framework for probing language models for social biases by assessing disparate treatment.
We curate SoFa, a large-scale benchmark designed to address the limitations of existing fairness collections.
We show that biases within language models are more nuanced than acknowledged, indicating a broader scope of encoded biases than previously recognized.
arXiv Detail & Related papers (2023-11-15T16:35:59Z) - Survey of Social Bias in Vision-Language Models [65.44579542312489]
Survey aims to provide researchers with a high-level insight into the similarities and differences of social bias studies in pre-trained models across NLP, CV, and VL.
The findings and recommendations presented here can benefit the ML community, fostering the development of fairer and non-biased AI models.
arXiv Detail & Related papers (2023-09-24T15:34:56Z) - Automated Ableism: An Exploration of Explicit Disability Biases in
Sentiment and Toxicity Analysis Models [5.611973529240434]
We analyze sentiment analysis and toxicity detection models to detect the presence of explicit bias against people with disability (PWD)
We employ the bias identification framework of Perturbation Sensitivity Analysis to examine conversations related to PWD on social media platforms, specifically Twitter and Reddit.
We then create the textitBias Identification Test in Sentiment (BITS) corpus to quantify explicit disability bias in any sentiment analysis and toxicity detection models.
arXiv Detail & Related papers (2023-07-18T12:45:54Z) - Bayesian Networks for the robust and unbiased prediction of depression
and its symptoms utilizing speech and multimodal data [65.28160163774274]
We apply a Bayesian framework to capture the relationships between depression, depression symptoms, and features derived from speech, facial expression and cognitive game data collected at thymia.
arXiv Detail & Related papers (2022-11-09T14:48:13Z) - A Survey of Methods for Addressing Class Imbalance in Deep-Learning
Based Natural Language Processing [68.37496795076203]
We provide guidance for NLP researchers and practitioners dealing with imbalanced data.
We first discuss various types of controlled and real-world class imbalance.
We organize the methods by whether they are based on sampling, data augmentation, choice of loss function, staged learning, or model design.
arXiv Detail & Related papers (2022-10-10T13:26:40Z) - Identification of Bias Against People with Disabilities in Sentiment
Analysis and Toxicity Detection Models [0.5758109624133713]
We present the Bias Identification Test in Sentiments (BITS), a corpus of 1,126 sentences designed to probe sentiment analysis models for biases in disability.
Results show that all exhibit strong negative biases on sentences that mention disability.
arXiv Detail & Related papers (2021-11-25T21:44:18Z) - Unpacking the Interdependent Systems of Discrimination: Ableist Bias in
NLP Systems through an Intersectional Lens [20.35460711907179]
We report on various analyses based on word predictions of a large-scale BERT language model.
Statistically significant results demonstrate that people with disabilities can be disadvantaged.
Findings also explore overlapping forms of discrimination related to interconnected gender and race identities.
arXiv Detail & Related papers (2021-10-01T16:40:58Z) - Towards Understanding and Mitigating Social Biases in Language Models [107.82654101403264]
Large-scale pretrained language models (LMs) can be potentially dangerous in manifesting undesirable representational biases.
We propose steps towards mitigating social biases during text generation.
Our empirical results and human evaluation demonstrate effectiveness in mitigating bias while retaining crucial contextual information.
arXiv Detail & Related papers (2021-06-24T17:52:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.