Towards classification parity across cohorts
- URL: http://arxiv.org/abs/2005.08033v1
- Date: Sat, 16 May 2020 16:31:08 GMT
- Title: Towards classification parity across cohorts
- Authors: Aarsh Patel, Rahul Gupta, Mukund Harakere, Satyapriya Krishna, Aman
Alok, Peng Liu
- Abstract summary: This research work aims to achieve classification parity across explicit as well as implicit sensitive features.
We obtain implicit cohorts by clustering embeddings of each individual trained on the language generated by them using a language model.
We improve classification parity by introducing modification to the loss function aimed to minimize the range of model performances across cohorts.
- Score: 16.21248370949611
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recently, there has been a lot of interest in ensuring algorithmic fairness
in machine learning where the central question is how to prevent sensitive
information (e.g. knowledge about the ethnic group of an individual) from
adding "unfair" bias to a learning algorithm (Feldman et al. (2015), Zemel et
al. (2013)). This has led to several debiasing algorithms on word embeddings
(Qian et al. (2019) , Bolukbasi et al. (2016)), coreference resolution (Zhao et
al. (2018a)), semantic role labeling (Zhao et al. (2017)), etc. Most of these
existing work deals with explicit sensitive features such as gender,
occupations or race which doesn't work with data where such features are not
captured due to privacy concerns. In this research work, we aim to achieve
classification parity across explicit as well as implicit sensitive features.
We define explicit cohorts as groups of people based on explicit sensitive
attributes provided in the data (age, gender, race) whereas implicit cohorts
are defined as groups of people with similar language usage. We obtain implicit
cohorts by clustering embeddings of each individual trained on the language
generated by them using a language model. We achieve two primary objectives in
this work : [1.] We experimented and discovered classification performance
differences across cohorts based on implicit and explicit features , [2] We
improved classification parity by introducing modification to the loss function
aimed to minimize the range of model performances across cohorts.
Related papers
- Minimax Group Fairness in Strategic Classification [8.250258160056514]
In strategic classification, agents manipulate their features, at a cost, to receive a positive classification outcome from the learner's classifier.
We consider learning objectives that have group fairness guarantees in addition to accuracy guarantees.
We formalize a fairness-aware Stackelberg game between a population of agents consisting of several groups, with each group having its own cost function.
arXiv Detail & Related papers (2024-10-03T14:22:55Z) - Evaluating the Fairness of Discriminative Foundation Models in Computer
Vision [51.176061115977774]
We propose a novel taxonomy for bias evaluation of discriminative foundation models, such as Contrastive Language-Pretraining (CLIP)
We then systematically evaluate existing methods for mitigating bias in these models with respect to our taxonomy.
Specifically, we evaluate OpenAI's CLIP and OpenCLIP models for key applications, such as zero-shot classification, image retrieval and image captioning.
arXiv Detail & Related papers (2023-10-18T10:32:39Z) - Human-Guided Fair Classification for Natural Language Processing [9.652938946631735]
We show how to leverage unsupervised style transfer and GPT-3's zero-shot capabilities to generate semantically similar sentences that differ along sensitive attributes.
We validate the generated pairs via an extensive crowdsourcing study, which confirms that a lot of these pairs align with human intuition about fairness in the context of toxicity classification.
arXiv Detail & Related papers (2022-12-20T10:46:40Z) - Fair Group-Shared Representations with Normalizing Flows [68.29997072804537]
We develop a fair representation learning algorithm which is able to map individuals belonging to different groups in a single group.
We show experimentally that our methodology is competitive with other fair representation learning algorithms.
arXiv Detail & Related papers (2022-01-17T10:49:49Z) - Social Norm Bias: Residual Harms of Fairness-Aware Algorithms [21.50551404445654]
Social Norm Bias (SNoB) is a subtle but consequential type of discrimination that may be exhibited by automated decision-making systems.
We quantify SNoB by measuring how an algorithm's predictions are associated with conformity to gender norms.
We show that post-processing interventions do not mitigate this type of bias at all.
arXiv Detail & Related papers (2021-08-25T05:54:56Z) - MultiFair: Multi-Group Fairness in Machine Learning [52.24956510371455]
We study multi-group fairness in machine learning (MultiFair)
We propose a generic end-to-end algorithmic framework to solve it.
Our proposed framework is generalizable to many different settings.
arXiv Detail & Related papers (2021-05-24T02:30:22Z) - Balancing Biases and Preserving Privacy on Balanced Faces in the Wild [50.915684171879036]
There are demographic biases present in current facial recognition (FR) models.
We introduce our Balanced Faces in the Wild dataset to measure these biases across different ethnic and gender subgroups.
We find that relying on a single score threshold to differentiate between genuine and imposters sample pairs leads to suboptimal results.
We propose a novel domain adaptation learning scheme that uses facial features extracted from state-of-the-art neural networks.
arXiv Detail & Related papers (2021-03-16T15:05:49Z) - Learning and Evaluating Representations for Deep One-class
Classification [59.095144932794646]
We present a two-stage framework for deep one-class classification.
We first learn self-supervised representations from one-class data, and then build one-class classifiers on learned representations.
In experiments, we demonstrate state-of-the-art performance on visual domain one-class classification benchmarks.
arXiv Detail & Related papers (2020-11-04T23:33:41Z) - Interpretable Assessment of Fairness During Model Evaluation [1.2183405753834562]
We introduce a novel hierarchical clustering algorithm to detect heterogeneity among users in given sets of sub-populations.
We demonstrate the performance of the algorithm on real data from LinkedIn.
arXiv Detail & Related papers (2020-10-26T02:31:17Z) - Weakly-Supervised Aspect-Based Sentiment Analysis via Joint
Aspect-Sentiment Topic Embedding [71.2260967797055]
We propose a weakly-supervised approach for aspect-based sentiment analysis.
We learn sentiment, aspect> joint topic embeddings in the word embedding space.
We then use neural models to generalize the word-level discriminative information.
arXiv Detail & Related papers (2020-10-13T21:33:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.