Human-Centric Multimodal Machine Learning: Recent Advances and Testbed
on AI-based Recruitment
- URL: http://arxiv.org/abs/2302.10908v1
- Date: Mon, 13 Feb 2023 16:44:44 GMT
- Title: Human-Centric Multimodal Machine Learning: Recent Advances and Testbed
on AI-based Recruitment
- Authors: Alejandro Pe\~na, Ignacio Serna, Aythami Morales, Julian Fierrez,
Alfonso Ortega, Ainhoa Herrarte, Manuel Alcantara and Javier Ortega-Garcia
- Abstract summary: There is a certain consensus about the need to develop AI applications with a Human-Centric approach.
Human-Centric Machine Learning needs to be developed based on four main requirements: (i) utility and social good; (ii) privacy and data ownership; (iii) transparency and accountability; and (iv) fairness in AI-driven decision-making processes.
We study how current multimodal algorithms based on heterogeneous sources of information are affected by sensitive elements and inner biases in the data.
- Score: 66.91538273487379
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The presence of decision-making algorithms in society is rapidly increasing
nowadays, while concerns about their transparency and the possibility of these
algorithms becoming new sources of discrimination are arising. There is a
certain consensus about the need to develop AI applications with a
Human-Centric approach. Human-Centric Machine Learning needs to be developed
based on four main requirements: (i) utility and social good; (ii) privacy and
data ownership; (iii) transparency and accountability; and (iv) fairness in
AI-driven decision-making processes. All these four Human-Centric requirements
are closely related to each other. With the aim of studying how current
multimodal algorithms based on heterogeneous sources of information are
affected by sensitive elements and inner biases in the data, we propose a
fictitious case study focused on automated recruitment: FairCVtest. We train
automatic recruitment algorithms using a set of multimodal synthetic profiles
including image, text, and structured data, which are consciously scored with
gender and racial biases. FairCVtest shows the capacity of the Artificial
Intelligence (AI) behind automatic recruitment tools built this way (a common
practice in many other application scenarios beyond recruitment) to extract
sensitive information from unstructured data and exploit it in combination to
data biases in undesirable (unfair) ways. We present an overview of recent
works developing techniques capable of removing sensitive information and
biases from the decision-making process of deep learning architectures, as well
as commonly used databases for fairness research in AI. We demonstrate how
learning approaches developed to guarantee privacy in latent spaces can lead to
unbiased and fair automatic decision-making process.
Related papers
- On Responsible Machine Learning Datasets with Fairness, Privacy, and Regulatory Norms [56.119374302685934]
There have been severe concerns over the trustworthiness of AI technologies.
Machine and deep learning algorithms depend heavily on the data used during their development.
We propose a framework to evaluate the datasets through a responsible rubric.
arXiv Detail & Related papers (2023-10-24T14:01:53Z) - What Matters in Learning from Offline Human Demonstrations for Robot
Manipulation [64.43440450794495]
We conduct an extensive study of six offline learning algorithms for robot manipulation.
Our study analyzes the most critical challenges when learning from offline human data.
We highlight opportunities for learning from human datasets.
arXiv Detail & Related papers (2021-08-06T20:48:30Z) - Towards Fairness Certification in Artificial Intelligence [31.920661197618195]
We propose a first joint effort to define the operational steps needed for AI fairness certification.
We will overview the criteria that should be met by an AI system before coming into official service and the conformity assessment procedures useful to monitor its functioning for fair decisions.
arXiv Detail & Related papers (2021-06-04T14:12:12Z) - Representative & Fair Synthetic Data [68.8204255655161]
We present a framework to incorporate fairness constraints into the self-supervised learning process.
We generate a representative as well as fair version of the UCI Adult census data set.
We consider representative & fair synthetic data a promising future building block to teach algorithms not on historic worlds, but rather on the worlds that we strive to live in.
arXiv Detail & Related papers (2021-04-07T09:19:46Z) - FairCVtest Demo: Understanding Bias in Multimodal Learning with a
Testbed in Fair Automatic Recruitment [79.23531577235887]
This demo shows the capacity of the Artificial Intelligence (AI) behind a recruitment tool to extract sensitive information from unstructured data.
Aditionally, the demo includes a new algorithm for discrimination-aware learning which eliminates sensitive information in our multimodal AI framework.
arXiv Detail & Related papers (2020-09-12T17:45:09Z) - Bias in Multimodal AI: Testbed for Fair Automatic Recruitment [73.85525896663371]
We study how current multimodal algorithms based on heterogeneous sources of information are affected by sensitive elements and inner biases in the data.
We train automatic recruitment algorithms using a set of multimodal synthetic profiles consciously scored with gender and racial biases.
Our methodology and results show how to generate fairer AI-based tools in general, and in particular fairer automated recruitment systems.
arXiv Detail & Related papers (2020-04-15T15:58:05Z) - FAE: A Fairness-Aware Ensemble Framework [18.993049769711114]
FAE (Fairness-Aware Ensemble) framework combines fairness-related interventions at both pre- and postprocessing steps of the data analysis process.
In the preprocessing step, we tackle the problems of under-representation of the protected group and of class-imbalance.
In the post-processing step, we tackle the problem of class overlapping by shifting the decision boundary in the direction of fairness.
arXiv Detail & Related papers (2020-02-03T13:05:18Z) - Bias in Data-driven AI Systems -- An Introductory Survey [37.34717604783343]
This survey focuses on data-driven AI, as a large part of AI is powered nowadays by (big) data and powerful Machine Learning (ML) algorithms.
If otherwise not specified, we use the general term bias to describe problems related to the gathering or processing of data that might result in prejudiced decisions on the bases of demographic features like race, sex, etc.
arXiv Detail & Related papers (2020-01-14T09:39:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.