Preserving Fairness in AI under Domain Shift
- URL: http://arxiv.org/abs/2301.12369v1
- Date: Sun, 29 Jan 2023 06:13:40 GMT
- Title: Preserving Fairness in AI under Domain Shift
- Authors: Serban Stan and Mohammad Rostami
- Abstract summary: Existing algorithms for ensuring fairness in AI use a single-shot training strategy.
We develop an algorithm to adapt a fair model to remain fair under domain shift.
- Score: 15.820660013260584
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Existing algorithms for ensuring fairness in AI use a single-shot training
strategy, where an AI model is trained on an annotated training dataset with
sensitive attributes and then fielded for utilization. This training strategy
is effective in problems with stationary distributions, where both training and
testing data are drawn from the same distribution. However, it is vulnerable
with respect to distributional shifts in the input space that may occur after
the initial training phase. As a result, the time-dependent nature of data can
introduce biases into the model predictions. Model retraining from scratch
using a new annotated dataset is a naive solution that is expensive and
time-consuming. We develop an algorithm to adapt a fair model to remain fair
under domain shift using solely new unannotated data points. We recast this
learning setting as an unsupervised domain adaptation problem. Our algorithm is
based on updating the model such that the internal representation of data
remains unbiased despite distributional shifts in the input space. We provide
extensive empirical validation on three widely employed fairness datasets to
demonstrate the effectiveness of our algorithm.
Related papers
- Continuous Unsupervised Domain Adaptation Using Stabilized
Representations and Experience Replay [23.871860648919593]
We introduce an algorithm for tackling the problem of unsupervised domain adaptation (UDA) in continual learning (CL) scenarios.
Our solution is based on stabilizing the learned internal distribution to enhances the model generalization on new domains.
We leverage experience replay to overcome the problem of catastrophic forgetting, where the model loses previously acquired knowledge when learning new tasks.
arXiv Detail & Related papers (2024-01-31T05:09:14Z) - Fairness and Accuracy under Domain Generalization [10.661409428935494]
Concerns have arisen that machine learning algorithms may be biased against certain social groups.
Many approaches have been proposed to make ML models fair, but they typically rely on the assumption that data distributions in training and deployment are identical.
We study the transfer of both fairness and accuracy under domain generalization where the data at test time may be sampled from never-before-seen domains.
arXiv Detail & Related papers (2023-01-30T23:10:17Z) - CAFA: Class-Aware Feature Alignment for Test-Time Adaptation [50.26963784271912]
Test-time adaptation (TTA) aims to address this challenge by adapting a model to unlabeled data at test time.
We propose a simple yet effective feature alignment loss, termed as Class-Aware Feature Alignment (CAFA), which simultaneously encourages a model to learn target representations in a class-discriminative manner.
arXiv Detail & Related papers (2022-06-01T03:02:07Z) - Leveraging Unlabeled Data to Predict Out-of-Distribution Performance [63.740181251997306]
Real-world machine learning deployments are characterized by mismatches between the source (training) and target (test) distributions.
In this work, we investigate methods for predicting the target domain accuracy using only labeled source data and unlabeled target data.
We propose Average Thresholded Confidence (ATC), a practical method that learns a threshold on the model's confidence, predicting accuracy as the fraction of unlabeled examples.
arXiv Detail & Related papers (2022-01-11T23:01:12Z) - BiFair: Training Fair Models with Bilevel Optimization [8.2509884277533]
We develop a new training algorithm, named BiFair, which jointly minimizes for a utility, and a fairness loss of interest.
Our algorithm consistently performs better, i.e., we reach to better values of a given fairness metric under same, or higher accuracy.
arXiv Detail & Related papers (2021-06-03T22:36:17Z) - Can Active Learning Preemptively Mitigate Fairness Issues? [66.84854430781097]
dataset bias is one of the prevailing causes of unfairness in machine learning.
We study whether models trained with uncertainty-based ALs are fairer in their decisions with respect to a protected class.
We also explore the interaction of algorithmic fairness methods such as gradient reversal (GRAD) and BALD.
arXiv Detail & Related papers (2021-04-14T14:20:22Z) - Adaptive Risk Minimization: Learning to Adapt to Domain Shift [109.87561509436016]
A fundamental assumption of most machine learning algorithms is that the training and test data are drawn from the same underlying distribution.
In this work, we consider the problem setting of domain generalization, where the training data are structured into domains and there may be multiple test time shifts.
We introduce the framework of adaptive risk minimization (ARM), in which models are directly optimized for effective adaptation to shift by learning to adapt on the training domains.
arXiv Detail & Related papers (2020-07-06T17:59:30Z) - Understanding Self-Training for Gradual Domain Adaptation [107.37869221297687]
We consider gradual domain adaptation, where the goal is to adapt an initial classifier trained on a source domain given only unlabeled data that shifts gradually in distribution towards a target domain.
We prove the first non-vacuous upper bound on the error of self-training with gradual shifts, under settings where directly adapting to the target domain can result in unbounded error.
The theoretical analysis leads to algorithmic insights, highlighting that regularization and label sharpening are essential even when we have infinite data, and suggesting that self-training works particularly well for shifts with small Wasserstein-infinity distance.
arXiv Detail & Related papers (2020-02-26T08:59:40Z) - Do We Really Need to Access the Source Data? Source Hypothesis Transfer
for Unsupervised Domain Adaptation [102.67010690592011]
Unsupervised adaptationUDA (UDA) aims to leverage the knowledge learned from a labeled source dataset to solve similar tasks in a new unlabeled domain.
Prior UDA methods typically require to access the source data when learning to adapt the model.
This work tackles a practical setting where only a trained source model is available and how we can effectively utilize such a model without source data to solve UDA problems.
arXiv Detail & Related papers (2020-02-20T03:13:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.