Social Biases in NLP Models as Barriers for Persons with Disabilities
- URL: http://arxiv.org/abs/2005.00813v1
- Date: Sat, 2 May 2020 12:16:54 GMT
- Title: Social Biases in NLP Models as Barriers for Persons with Disabilities
- Authors: Ben Hutchinson, Vinodkumar Prabhakaran, Emily Denton, Kellie Webster,
Yu Zhong, Stephen Denuyl
- Abstract summary: We present evidence of undesirable biases towards mentions of disability in two different English language models: toxicity prediction and sentiment analysis.
Next, we demonstrate that the neural embeddings that are the critical first step in most NLP pipelines similarly contain undesirable biases towards mentions of disability.
We end by highlighting topical biases in the discourse about disability which may contribute to the observed model biases.
- Score: 13.579848462349192
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Building equitable and inclusive NLP technologies demands consideration of
whether and how social attitudes are represented in ML models. In particular,
representations encoded in models often inadvertently perpetuate undesirable
social biases from the data on which they are trained. In this paper, we
present evidence of such undesirable biases towards mentions of disability in
two different English language models: toxicity prediction and sentiment
analysis. Next, we demonstrate that the neural embeddings that are the critical
first step in most NLP pipelines similarly contain undesirable biases towards
mentions of disability. We end by highlighting topical biases in the discourse
about disability which may contribute to the observed model biases; for
instance, gun violence, homelessness, and drug addiction are over-represented
in texts discussing mental illness.
Related papers
- Interpretable Debiasing of Vision-Language Models for Social Fairness [55.85977929985967]
We introduce an interpretable, model-agnostic bias mitigation framework, DeBiasLens, that localizes social attribute neurons in Vision-Language models.<n>We train SAEs on facial image or caption datasets without corresponding social attribute labels to uncover neurons highly responsive to specific demographics.<n>Our research lays the groundwork for future auditing tools, prioritizing social fairness in emerging real-world AI systems.
arXiv Detail & Related papers (2026-02-27T13:37:11Z) - Auditing Disability Representation in Vision-Language Models [0.6987503477818553]
We study disability aware descriptions for person centric images.<n>We introduce a benchmark based on paired Neutral Prompts (NP) and Disability-Contextualised Prompts (DP)<n>We evaluate 15 state-of-the-art open- and closed-source vision-language models under a zero-shot setting across 9 disability categories.
arXiv Detail & Related papers (2026-01-24T07:25:43Z) - SocialNLI: A Dialogue-Centric Social Inference Dataset [49.60157928163403]
We introduce SocialNLI -- the first social dialogue inference dataset.<n>SocialNLI consists of a collection of dialogue transcripts hand-picked to center complex social nuances.<n>We evaluate reasoning models theory-of-mind ability through multi-step counterfactual reasoning.
arXiv Detail & Related papers (2025-10-06T23:42:01Z) - Who's Asking? Investigating Bias Through the Lens of Disability Framed Queries in LLMs [2.722784054643991]
Large Language Models (LLMs) routinely infer users demographic traits from phrasing alone.<n>Disability cues in shaping these inferences remains largely uncharted.<n>We present the first systematic audit of disability-conditioned demographic bias across eight state-of-the-art instruction-tuned LLMs.
arXiv Detail & Related papers (2025-08-18T21:03:09Z) - Misspellings in Natural Language Processing: A survey [52.419589623702336]
misspellings have become ubiquitous in digital communication.
We reconstruct a history of misspellings as a scientific problem.
We discuss the latest advancements to address the challenge of misspellings in NLP.
arXiv Detail & Related papers (2025-01-28T10:26:04Z) - The Devil is in the Neurons: Interpreting and Mitigating Social Biases in Pre-trained Language Models [78.69526166193236]
Pre-trained Language models (PLMs) have been acknowledged to contain harmful information, such as social biases.
We propose sc Social Bias Neurons to accurately pinpoint units (i.e., neurons) in a language model that can be attributed to undesirable behavior, such as social bias.
As measured by prior metrics from StereoSet, our model achieves a higher degree of fairness while maintaining language modeling ability with low cost.
arXiv Detail & Related papers (2024-06-14T15:41:06Z) - Decoding Susceptibility: Modeling Misbelief to Misinformation Through a Computational Approach [61.04606493712002]
Susceptibility to misinformation describes the degree of belief in unverifiable claims that is not observable.
Existing susceptibility studies heavily rely on self-reported beliefs.
We propose a computational approach to model users' latent susceptibility levels.
arXiv Detail & Related papers (2023-11-16T07:22:56Z) - Social Bias Probing: Fairness Benchmarking for Language Models [38.180696489079985]
This paper proposes a novel framework for probing language models for social biases by assessing disparate treatment.
We curate SoFa, a large-scale benchmark designed to address the limitations of existing fairness collections.
We show that biases within language models are more nuanced than acknowledged, indicating a broader scope of encoded biases than previously recognized.
arXiv Detail & Related papers (2023-11-15T16:35:59Z) - Survey of Social Bias in Vision-Language Models [65.44579542312489]
Survey aims to provide researchers with a high-level insight into the similarities and differences of social bias studies in pre-trained models across NLP, CV, and VL.
The findings and recommendations presented here can benefit the ML community, fostering the development of fairer and non-biased AI models.
arXiv Detail & Related papers (2023-09-24T15:34:56Z) - Automated Ableism: An Exploration of Explicit Disability Biases in
Sentiment and Toxicity Analysis Models [5.611973529240434]
We analyze sentiment analysis and toxicity detection models to detect the presence of explicit bias against people with disability (PWD)
We employ the bias identification framework of Perturbation Sensitivity Analysis to examine conversations related to PWD on social media platforms, specifically Twitter and Reddit.
We then create the textitBias Identification Test in Sentiment (BITS) corpus to quantify explicit disability bias in any sentiment analysis and toxicity detection models.
arXiv Detail & Related papers (2023-07-18T12:45:54Z) - Bayesian Networks for the robust and unbiased prediction of depression
and its symptoms utilizing speech and multimodal data [65.28160163774274]
We apply a Bayesian framework to capture the relationships between depression, depression symptoms, and features derived from speech, facial expression and cognitive game data collected at thymia.
arXiv Detail & Related papers (2022-11-09T14:48:13Z) - A Survey of Methods for Addressing Class Imbalance in Deep-Learning
Based Natural Language Processing [68.37496795076203]
We provide guidance for NLP researchers and practitioners dealing with imbalanced data.
We first discuss various types of controlled and real-world class imbalance.
We organize the methods by whether they are based on sampling, data augmentation, choice of loss function, staged learning, or model design.
arXiv Detail & Related papers (2022-10-10T13:26:40Z) - Identification of Bias Against People with Disabilities in Sentiment
Analysis and Toxicity Detection Models [0.5758109624133713]
We present the Bias Identification Test in Sentiments (BITS), a corpus of 1,126 sentences designed to probe sentiment analysis models for biases in disability.
Results show that all exhibit strong negative biases on sentences that mention disability.
arXiv Detail & Related papers (2021-11-25T21:44:18Z) - Unpacking the Interdependent Systems of Discrimination: Ableist Bias in
NLP Systems through an Intersectional Lens [20.35460711907179]
We report on various analyses based on word predictions of a large-scale BERT language model.
Statistically significant results demonstrate that people with disabilities can be disadvantaged.
Findings also explore overlapping forms of discrimination related to interconnected gender and race identities.
arXiv Detail & Related papers (2021-10-01T16:40:58Z) - Towards Understanding and Mitigating Social Biases in Language Models [107.82654101403264]
Large-scale pretrained language models (LMs) can be potentially dangerous in manifesting undesirable representational biases.
We propose steps towards mitigating social biases during text generation.
Our empirical results and human evaluation demonstrate effectiveness in mitigating bias while retaining crucial contextual information.
arXiv Detail & Related papers (2021-06-24T17:52:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.