National Origin Discrimination in Deep-learning-powered Automated Resume
Screening
- URL: http://arxiv.org/abs/2307.08624v1
- Date: Thu, 13 Jul 2023 01:35:29 GMT
- Title: National Origin Discrimination in Deep-learning-powered Automated Resume
Screening
- Authors: Sihang Li, Kuangzheng Li, Haibing Lu
- Abstract summary: Many companies and organizations have started to use some form of AIenabled auto mated tools to assist in their hiring process.
There are increasing concerns on unfair treatment to candidates, caused by underlying bias in AI systems.
This study examined deep learning methods, a recent technology breakthrough, with focus on their application to automated resume screening.
- Score: 3.251347385432286
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Many companies and organizations have started to use some form of AIenabled
auto mated tools to assist in their hiring process, e.g. screening resumes,
interviewing candi dates, performance evaluation. While those AI tools have
greatly improved human re source operations efficiency and provided
conveniences to job seekers as well, there are increasing concerns on unfair
treatment to candidates, caused by underlying bias in AI systems. Laws around
equal opportunity and fairness, like GDPR, CCPA, are introduced or under
development, in attempt to regulate AI. However, it is difficult to implement
AI regulations in practice, as technologies are constantly advancing and the
risk perti nent to their applications can fail to be recognized. This study
examined deep learning methods, a recent technology breakthrough, with focus on
their application to automated resume screening. One impressive performance of
deep learning methods is the represen tation of individual words as
lowdimensional numerical vectors, called word embedding, which are learned from
aggregated global wordword cooccurrence statistics from a cor pus, like
Wikipedia or Google news. The resulting word representations possess interest
ing linear substructures of the word vector space and have been widely used in
down stream tasks, like resume screening. However, word embedding inherits and
reinforces the stereotyping from the training corpus, as deep learning models
essentially learn a probability distribution of words and their relations from
history data. Our study finds out that if we rely on such deeplearningpowered
automated resume screening tools, it may lead to decisions favoring or
disfavoring certain demographic groups and raise eth ical, even legal,
concerns. To address the issue, we developed bias mitigation method. Extensive
experiments on real candidate resumes are conducted to validate our study
Related papers
- Generative AI in Writing Research Papers: A New Type of Algorithmic Bias
and Uncertainty in Scholarly Work [0.38850145898707145]
Large language models (LLMs) and generative AI tools present challenges in identifying and addressing biases.
generative AI tools are susceptible to goal misgeneralization, hallucinations, and adversarial attacks such as red teaming prompts.
We find that incorporating generative AI in the process of writing research manuscripts introduces a new type of context-induced algorithmic bias.
arXiv Detail & Related papers (2023-12-04T04:05:04Z) - Combatting Human Trafficking in the Cyberspace: A Natural Language
Processing-Based Methodology to Analyze the Language in Online Advertisements [55.2480439325792]
This project tackles the pressing issue of human trafficking in online C2C marketplaces through advanced Natural Language Processing (NLP) techniques.
We introduce a novel methodology for generating pseudo-labeled datasets with minimal supervision, serving as a rich resource for training state-of-the-art NLP models.
A key contribution is the implementation of an interpretability framework using Integrated Gradients, providing explainable insights crucial for law enforcement.
arXiv Detail & Related papers (2023-11-22T02:45:01Z) - Unsupervised Sentiment Analysis of Plastic Surgery Social Media Posts [91.3755431537592]
The massive collection of user posts across social media platforms is primarily untapped for artificial intelligence (AI) use cases.
Natural language processing (NLP) is a subfield of AI that leverages bodies of documents, known as corpora, to train computers in human-like language understanding.
This study demonstrates that the applied results of unsupervised analysis allow a computer to predict either negative, positive, or neutral user sentiment towards plastic surgery.
arXiv Detail & Related papers (2023-07-05T20:16:20Z) - Implementing Active Learning in Cybersecurity: Detecting Anomalies in
Redacted Emails [10.303697869042283]
We present research results concerning the application of Active Learning to anomaly detection in redacted emails.
We evaluate different AL strategies and their impact on resulting model performance.
arXiv Detail & Related papers (2023-03-01T23:53:01Z) - Human-Centric Multimodal Machine Learning: Recent Advances and Testbed
on AI-based Recruitment [66.91538273487379]
There is a certain consensus about the need to develop AI applications with a Human-Centric approach.
Human-Centric Machine Learning needs to be developed based on four main requirements: (i) utility and social good; (ii) privacy and data ownership; (iii) transparency and accountability; and (iv) fairness in AI-driven decision-making processes.
We study how current multimodal algorithms based on heterogeneous sources of information are affected by sensitive elements and inner biases in the data.
arXiv Detail & Related papers (2023-02-13T16:44:44Z) - A methodology to characterize bias and harmful stereotypes in natural
language processing in Latin America [2.05094736006609]
We show how social scientists, domain experts, and machine learning experts can collaboratively explore biases and harmful stereotypes in word embeddings and large language models.
Our methodology is based on the following principles.
arXiv Detail & Related papers (2022-07-14T01:07:55Z) - Verifying Learning-Based Robotic Navigation Systems [61.01217374879221]
We show how modern verification engines can be used for effective model selection.
Specifically, we use verification to detect and rule out policies that may demonstrate suboptimal behavior.
Our work is the first to demonstrate the use of verification backends for recognizing suboptimal DRL policies in real-world robots.
arXiv Detail & Related papers (2022-05-26T17:56:43Z) - FairCVtest Demo: Understanding Bias in Multimodal Learning with a
Testbed in Fair Automatic Recruitment [79.23531577235887]
This demo shows the capacity of the Artificial Intelligence (AI) behind a recruitment tool to extract sensitive information from unstructured data.
Aditionally, the demo includes a new algorithm for discrimination-aware learning which eliminates sensitive information in our multimodal AI framework.
arXiv Detail & Related papers (2020-09-12T17:45:09Z) - Bias in Multimodal AI: Testbed for Fair Automatic Recruitment [73.85525896663371]
We study how current multimodal algorithms based on heterogeneous sources of information are affected by sensitive elements and inner biases in the data.
We train automatic recruitment algorithms using a set of multimodal synthetic profiles consciously scored with gender and racial biases.
Our methodology and results show how to generate fairer AI-based tools in general, and in particular fairer automated recruitment systems.
arXiv Detail & Related papers (2020-04-15T15:58:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.