Context-Aware Discrimination Detection in Job Vacancies using
Computational Language Models
- URL: http://arxiv.org/abs/2202.03907v1
- Date: Wed, 2 Feb 2022 09:25:08 GMT
- Title: Context-Aware Discrimination Detection in Job Vacancies using
Computational Language Models
- Authors: S. Vethman, A. Adhikari, M. H. T. de Boer, J. A. G. M. van Genabeek,
C. J. Veenman
- Abstract summary: Discriminatory job vacancies are disapproved worldwide, but remain persistent.
Discriminatory job vacancies can be explicit by directly referring to demographic memberships of candidates.
implicit forms of discrimination are also present that may not always be illegal but still influence the diversity of applicants.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Discriminatory job vacancies are disapproved worldwide, but remain
persistent. Discrimination in job vacancies can be explicit by directly
referring to demographic memberships of candidates. More implicit forms of
discrimination are also present that may not always be illegal but still
influence the diversity of applicants. Explicit written discrimination is still
present in numerous job vacancies, as was recently observed in the Netherlands.
Current efforts for the detection of explicit discrimination concern the
identification of job vacancies containing potentially discriminating terms
such as "young" or "male". However, automatic detection is inefficient due to
low precision: e.g. "we are a young company" or "working with mostly male
patients" are phrases that contain explicit terms, while the context shows that
these do not reflect discriminatory content.
In this paper, we show how machine learning based computational language
models can raise precision in the detection of explicit discrimination by
identifying when the potentially discriminating terms are used in a
discriminatory context. We focus on gender discrimination, which indeed suffers
from low precision when filtering explicit terms. First, we created a data set
for gender discrimination in job vacancies. Second, we investigated a variety
of computational language models for discriminatory context detection. Third,
we evaluated the capability of these models to detect unforeseen discriminating
terms in context. The results show that machine learning based methods can
detect explicit gender discrimination with high precision and help in finding
new forms of discrimination. Accordingly, the proposed methods can
substantially increase the effectiveness of detecting job vacancies which are
highly suspected to be discriminatory. In turn, this may lower the
discrimination experienced at the start of the recruitment process.
Related papers
- Beyond Binary Gender: Evaluating Gender-Inclusive Machine Translation with Ambiguous Attitude Words [85.48043537327258]
Existing machine translation gender bias evaluations are primarily focused on male and female genders.
This study presents a benchmark AmbGIMT (Gender-Inclusive Machine Translation with Ambiguous attitude words)
We propose a novel process to evaluate gender bias based on the Emotional Attitude Score (EAS), which is used to quantify ambiguous attitude words.
arXiv Detail & Related papers (2024-07-23T08:13:51Z) - Auditing for Racial Discrimination in the Delivery of Education Ads [50.37313459134418]
We propose a new third-party auditing method that can evaluate racial bias in the delivery of ads for education opportunities.
We find evidence of racial discrimination in Meta's algorithmic delivery of ads for education opportunities, posing legal and ethical concerns.
arXiv Detail & Related papers (2024-06-02T02:00:55Z) - Unlawful Proxy Discrimination: A Framework for Challenging Inherently Discriminatory Algorithms [4.1221687771754]
EU legal concept of direct discrimination may apply to various algorithmic decision-making contexts.
Unlike indirect discrimination, there is generally no 'objective justification' stage in the direct discrimination framework.
We focus on the most likely candidate for direct discrimination in the algorithmic context.
arXiv Detail & Related papers (2024-04-22T10:06:17Z) - Generalizing Fairness to Generative Language Models via Reformulation of Non-discrimination Criteria [4.738231680800414]
This paper studies how to uncover and quantify the presence of gender biases in generative language models.
We derive generative AI analogues of three well-known non-discrimination criteria from classification, namely independence, separation and sufficiency.
Our results address the presence of occupational gender bias within such conversational language models.
arXiv Detail & Related papers (2024-03-13T14:19:08Z) - Discrimination through Image Selection by Job Advertisers on Facebook [79.21648699199648]
We propose and investigate the prevalence of a new means for discrimination in job advertising.
It combines both targeting and delivery -- through the disproportionate representation or exclusion of people of certain demographics in job ad images.
We use the Facebook Ad Library to demonstrate the prevalence of this practice.
arXiv Detail & Related papers (2023-06-13T03:43:58Z) - Reusing the Task-specific Classifier as a Discriminator:
Discriminator-free Adversarial Domain Adaptation [55.27563366506407]
We introduce a discriminator-free adversarial learning network (DALN) for unsupervised domain adaptation (UDA)
DALN achieves explicit domain alignment and category distinguishment through a unified objective.
DALN compares favorably against the existing state-of-the-art (SOTA) methods on a variety of public datasets.
arXiv Detail & Related papers (2022-04-08T04:40:18Z) - Statistical discrimination in learning agents [64.78141757063142]
Statistical discrimination emerges in agent policies as a function of both the bias in the training population and of agent architecture.
We show that less discrimination emerges with agents that use recurrent neural networks, and when their training environment has less bias.
arXiv Detail & Related papers (2021-10-21T18:28:57Z) - Training GANs with Stronger Augmentations via Contrastive Discriminator [80.8216679195]
We introduce a contrastive representation learning scheme into the GAN discriminator, coined ContraD.
This "fusion" enables the discriminators to work with much stronger augmentations without increasing their training instability.
Our experimental results show that GANs with ContraD consistently improve FID and IS compared to other recent techniques incorporating data augmentations.
arXiv Detail & Related papers (2021-03-17T16:04:54Z) - A Normative approach to Attest Digital Discrimination [6.372554934045607]
Examples include low-income neighbourhood's targeted with high-interest loans or low credit scores, and women being undervalued by 21% in online marketing.
We use norms as an abstraction to represent different situations that may lead to digital discrimination.
In particular, we formalise non-discrimination norms in the context of ML systems and propose an algorithm to check whether ML systems violate these norms.
arXiv Detail & Related papers (2020-07-14T15:14:52Z) - Why Fairness Cannot Be Automated: Bridging the Gap Between EU
Non-Discrimination Law and AI [10.281644134255576]
Article identifies a critical incompatibility between European notions of discrimination and existing statistical measures of fairness.
We show how the legal protection offered by non-discrimination law is challenged when AI, not humans, discriminate.
We propose "conditional demographic disparity" (CDD) as a standard baseline statistical measurement.
arXiv Detail & Related papers (2020-05-12T16:30:12Z) - Demographics Should Not Be the Reason of Toxicity: Mitigating
Discrimination in Text Classifications with Instance Weighting [36.87473475196733]
We formalize the unintended biases in text classification datasets as a kind of selection bias from the non-discrimination distribution to the discrimination distribution.
Our method can effectively alleviate the impacts of the unintended biases without significantly hurting models' generalization ability.
arXiv Detail & Related papers (2020-04-29T11:22:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.