On the Origins of Bias in NLP through the Lens of the Jim Code
- URL: http://arxiv.org/abs/2305.09281v1
- Date: Tue, 16 May 2023 08:37:13 GMT
- Title: On the Origins of Bias in NLP through the Lens of the Jim Code
- Authors: Fatma Elsafoury, Gavin Abercrombie
- Abstract summary: We trace the biases in current natural language processing (NLP) models back to their origins in racism, sexism, and homophobia over the last 500 years.
We show how the causes of the biases in the NLP pipeline are rooted in social issues.
- Score: 1.256413718364189
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we trace the biases in current natural language processing
(NLP) models back to their origins in racism, sexism, and homophobia over the
last 500 years. We review literature from critical race theory, gender studies,
data ethics, and digital humanities studies, and summarize the origins of bias
in NLP models from these social science perspective. We show how the causes of
the biases in the NLP pipeline are rooted in social issues. Finally, we argue
that the only way to fix the bias and unfairness in NLP is by addressing the
social problems that caused them in the first place and by incorporating social
sciences and social scientists in efforts to mitigate bias in NLP models. We
provide actionable recommendations for the NLP research community to do so.
Related papers
- Exploring the Jungle of Bias: Political Bias Attribution in Language Models via Dependency Analysis [86.49858739347412]
Large Language Models (LLMs) have sparked intense debate regarding the prevalence of bias in these models and its mitigation.
We propose a prompt-based method for the extraction of confounding and mediating attributes which contribute to the decision process.
We find that the observed disparate treatment can at least in part be attributed to confounding and mitigating attributes and model misalignment.
arXiv Detail & Related papers (2023-11-15T00:02:25Z) - A Material Lens on Coloniality in NLP [57.63027898794855]
Coloniality is the continuation of colonial harms beyond "official" colonization.
We argue that coloniality is implicitly embedded in and amplified by NLP data, algorithms, and software.
arXiv Detail & Related papers (2023-11-14T18:52:09Z) - Thesis Distillation: Investigating The Impact of Bias in NLP Models on
Hate Speech Detection [6.2548734896918505]
This paper is a summary of the work done in my PhD thesis.
I investigate the impact of bias in NLP models on the task of hate speech detection from three perspectives.
arXiv Detail & Related papers (2023-08-31T08:40:41Z) - Beyond Good Intentions: Reporting the Research Landscape of NLP for
Social Good [115.1507728564964]
We introduce NLP4SG Papers, a scientific dataset with three associated tasks.
These tasks help identify NLP4SG papers and characterize the NLP4SG landscape.
We use state-of-the-art NLP models to address each of these tasks and use them on the entire ACL Anthology.
arXiv Detail & Related papers (2023-05-09T14:16:25Z) - A Survey of Methods for Addressing Class Imbalance in Deep-Learning
Based Natural Language Processing [68.37496795076203]
We provide guidance for NLP researchers and practitioners dealing with imbalanced data.
We first discuss various types of controlled and real-world class imbalance.
We organize the methods by whether they are based on sampling, data augmentation, choice of loss function, staged learning, or model design.
arXiv Detail & Related papers (2022-10-10T13:26:40Z) - Towards an Enhanced Understanding of Bias in Pre-trained Neural Language
Models: A Survey with Special Emphasis on Affective Bias [2.6304695993930594]
We present a survey to comprehend bias in large pre-trained language models, analyze the stages at which they occur, and various ways in which these biases could be quantified and mitigated.
Considering wide applicability of textual affective computing based downstream tasks in real-world systems such as business, healthcare, education, etc., we give a special emphasis on investigating bias in the context of affect (emotion) i.e., Affective Bias.
We present a summary of various bias evaluation corpora that help to aid future research and discuss challenges in the research on bias in pre-trained language models.
arXiv Detail & Related papers (2022-04-21T18:51:19Z) - A Survey on Bias and Fairness in Natural Language Processing [1.713291434132985]
We analyze the origins of biases, the definitions of fairness, and how different subfields of NLP bias can be mitigated.
We discuss how future studies can work towards eradicating pernicious biases from NLP algorithms.
arXiv Detail & Related papers (2022-03-06T18:12:30Z) - Argument from Old Man's View: Assessing Social Bias in Argumentation [20.65183968971417]
Social bias in language poses a problem with ethical impact for many NLP applications.
Recent research has shown that machine learning models trained on respective data may not only adopt, but even amplify the bias.
We study the existence of social biases in large English debate portals.
arXiv Detail & Related papers (2020-11-24T10:39:44Z) - Case Study: Deontological Ethics in NLP [119.53038547411062]
We study one ethical theory, namely deontological ethics, from the perspective of NLP.
In particular, we focus on the generalization principle and the respect for autonomy through informed consent.
We provide four case studies to demonstrate how these principles can be used with NLP systems.
arXiv Detail & Related papers (2020-10-09T16:04:51Z) - Towards Debiasing Sentence Representations [109.70181221796469]
We show that Sent-Debias is effective in removing biases, and at the same time, preserves performance on sentence-level downstream tasks.
We hope that our work will inspire future research on characterizing and removing social biases from widely adopted sentence representations for fairer NLP.
arXiv Detail & Related papers (2020-07-16T04:22:30Z) - Language (Technology) is Power: A Critical Survey of "Bias" in NLP [11.221552724154986]
We survey 146 papers analyzing "bias" in NLP systems.
We find that their motivations are vague, inconsistent, and lacking in normative reasoning.
We propose three recommendations that should guide work analyzing "bias" in NLP systems.
arXiv Detail & Related papers (2020-05-28T14:32:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.