Discovering and Categorising Language Biases in Reddit
- URL: http://arxiv.org/abs/2008.02754v2
- Date: Thu, 13 Aug 2020 18:38:21 GMT
- Title: Discovering and Categorising Language Biases in Reddit
- Authors: Xavier Ferrer, Tom van Nuenen, Jose M. Such, Natalia Criado
- Abstract summary: This paper proposes a data-driven approach to automatically discover language biases encoded in the vocabulary of online discourse communities on Reddit.
We use word embeddings to transform text into high-dimensional dense vectors and capture semantic relations between words.
We successfully discover gender bias, religion bias, and ethnic bias in different Reddit communities.
- Score: 5.670038395203354
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a data-driven approach using word embeddings to discover and
categorise language biases on the discussion platform Reddit. As spaces for
isolated user communities, platforms such as Reddit are increasingly connected
to issues of racism, sexism and other forms of discrimination. Hence, there is
a need to monitor the language of these groups. One of the most promising AI
approaches to trace linguistic biases in large textual datasets involves word
embeddings, which transform text into high-dimensional dense vectors and
capture semantic relations between words. Yet, previous studies require
predefined sets of potential biases to study, e.g., whether gender is more or
less associated with particular types of jobs. This makes these approaches
unfit to deal with smaller and community-centric datasets such as those on
Reddit, which contain smaller vocabularies and slang, as well as biases that
may be particular to that community. This paper proposes a data-driven approach
to automatically discover language biases encoded in the vocabulary of online
discourse communities on Reddit. In our approach, protected attributes are
connected to evaluative words found in the data, which are then categorised
through a semantic analysis system. We verify the effectiveness of our method
by comparing the biases we discover in the Google News dataset with those found
in previous literature. We then successfully discover gender bias, religion
bias, and ethnic bias in different Reddit communities. We conclude by
discussing potential application scenarios and limitations of this data-driven
bias discovery method.
Related papers
- Global Voices, Local Biases: Socio-Cultural Prejudices across Languages [22.92083941222383]
Human biases are ubiquitous but not uniform; disparities exist across linguistic, cultural, and societal borders.
In this work, we scale the Word Embedding Association Test (WEAT) to 24 languages, enabling broader studies.
To encompass more widely prevalent societal biases, we examine new bias dimensions across toxicity, ableism, and more.
arXiv Detail & Related papers (2023-10-26T17:07:50Z) - Evaluating Biased Attitude Associations of Language Models in an
Intersectional Context [2.891314299138311]
Language models are trained on large-scale corpora that embed implicit biases documented in psychology.
We study biases related to age, education, gender, height, intelligence, literacy, race, religion, sex, sexual orientation, social class, and weight.
We find that language models exhibit the most biased attitudes against gender identity, social class, and sexual orientation signals in language.
arXiv Detail & Related papers (2023-07-07T03:01:56Z) - Detecting Unintended Social Bias in Toxic Language Datasets [32.724030288421474]
This paper introduces a new dataset ToxicBias curated from the existing dataset of Kaggle competition named "Jigsaw Unintended Bias in Toxicity Classification"
The dataset contains instances annotated for five different bias categories, viz., gender, race/ethnicity, religion, political, and LGBTQ.
We train transformer-based models using our curated datasets and report baseline performance for bias identification, target generation, and bias implications.
arXiv Detail & Related papers (2022-10-21T06:50:12Z) - A New Generation of Perspective API: Efficient Multilingual
Character-level Transformers [66.9176610388952]
We present the fundamentals behind the next version of the Perspective API from Google Jigsaw.
At the heart of the approach is a single multilingual token-free Charformer model.
We demonstrate that by forgoing static vocabularies, we gain flexibility across a variety of settings.
arXiv Detail & Related papers (2022-02-22T20:55:31Z) - WordBias: An Interactive Visual Tool for Discovering Intersectional
Biases Encoded in Word Embeddings [39.87681037622605]
We present WordBias, an interactive visual tool designed to explore biases against intersectional groups encoded in word embeddings.
Given a pretrained static word embedding, WordBias computes the association of each word along different groups based on race, age, etc.
arXiv Detail & Related papers (2021-03-05T11:04:35Z) - Discovering and Interpreting Biased Concepts in Online Communities [5.670038395203354]
Language carries implicit human biases, functioning both as a reflection and a perpetuation of stereotypes that people carry with them.
ML-based NLP methods such as word embeddings have been shown to learn such language biases with striking accuracy.
This paper improves upon, extends, and evaluates our previous data-driven method to automatically discover and help interpret biased concepts encoded in word embeddings.
arXiv Detail & Related papers (2020-10-27T17:07:12Z) - Towards Debiasing Sentence Representations [109.70181221796469]
We show that Sent-Debias is effective in removing biases, and at the same time, preserves performance on sentence-level downstream tasks.
We hope that our work will inspire future research on characterizing and removing social biases from widely adopted sentence representations for fairer NLP.
arXiv Detail & Related papers (2020-07-16T04:22:30Z) - Double-Hard Debias: Tailoring Word Embeddings for Gender Bias Mitigation [94.98656228690233]
We propose a technique that purifies the word embeddings against corpus regularities prior to inferring and removing the gender subspace.
Our approach preserves the distributional semantics of the pre-trained word embeddings while reducing gender bias to a significantly larger degree than prior approaches.
arXiv Detail & Related papers (2020-05-03T02:33:20Z) - Gender Bias in Multilingual Embeddings and Cross-Lingual Transfer [101.58431011820755]
We study gender bias in multilingual embeddings and how it affects transfer learning for NLP applications.
We create a multilingual dataset for bias analysis and propose several ways for quantifying bias in multilingual representations.
arXiv Detail & Related papers (2020-05-02T04:34:37Z) - Multi-Dimensional Gender Bias Classification [67.65551687580552]
Machine learning models can inadvertently learn socially undesirable patterns when training on gender biased text.
We propose a general framework that decomposes gender bias in text along several pragmatic and semantic dimensions.
Using this fine-grained framework, we automatically annotate eight large scale datasets with gender information.
arXiv Detail & Related papers (2020-05-01T21:23:20Z) - Towards Controllable Biases in Language Generation [87.89632038677912]
We develop a method to induce societal biases in generated text when input prompts contain mentions of specific demographic groups.
We analyze two scenarios: 1) inducing negative biases for one demographic and positive biases for another demographic, and 2) equalizing biases between demographics.
arXiv Detail & Related papers (2020-05-01T08:25:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.