Can Active Learning Preemptively Mitigate Fairness Issues?
- URL: http://arxiv.org/abs/2104.06879v1
- Date: Wed, 14 Apr 2021 14:20:22 GMT
- Title: Can Active Learning Preemptively Mitigate Fairness Issues?
- Authors: Fr\'ed\'eric Branchaud-Charron, Parmida Atighehchian, Pau Rodr\'iguez,
Grace Abuhamad, Alexandre Lacoste
- Abstract summary: dataset bias is one of the prevailing causes of unfairness in machine learning.
We study whether models trained with uncertainty-based ALs are fairer in their decisions with respect to a protected class.
We also explore the interaction of algorithmic fairness methods such as gradient reversal (GRAD) and BALD.
- Score: 66.84854430781097
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Dataset bias is one of the prevailing causes of unfairness in machine
learning. Addressing fairness at the data collection and dataset preparation
stages therefore becomes an essential part of training fairer algorithms. In
particular, active learning (AL) algorithms show promise for the task by
drawing importance to the most informative training samples. However, the
effect and interaction between existing AL algorithms and algorithmic fairness
remain under-explored. In this paper, we study whether models trained with
uncertainty-based AL heuristics such as BALD are fairer in their decisions with
respect to a protected class than those trained with identically independently
distributed (i.i.d.) sampling. We found a significant improvement on predictive
parity when using BALD, while also improving accuracy compared to i.i.d.
sampling. We also explore the interaction of algorithmic fairness methods such
as gradient reversal (GRAD) and BALD. We found that, while addressing different
fairness issues, their interaction further improves the results on most
benchmarks and metrics we explored.
Related papers
- Fair Class-Incremental Learning using Sample Weighting [27.82760149957115]
We show that naively using all the samples of the current task for training results in unfair catastrophic forgetting for certain sensitive groups including classes.
We propose a fair class-incremental learning framework that adjusts the training weights of current task samples to change the direction of the average gradient vector.
Experiments show that FSW achieves better accuracy-fairness tradeoff results than state-of-the-art approaches on real datasets.
arXiv Detail & Related papers (2024-10-02T08:32:21Z) - Fairness Without Harm: An Influence-Guided Active Sampling Approach [32.173195437797766]
We aim to train models that mitigate group fairness disparity without causing harm to model accuracy.
The current data acquisition methods, such as fair active learning approaches, typically require annotating sensitive attributes.
We propose a tractable active data sampling algorithm that does not rely on training group annotations.
arXiv Detail & Related papers (2024-02-20T07:57:38Z) - Noisy Correspondence Learning with Self-Reinforcing Errors Mitigation [63.180725016463974]
Cross-modal retrieval relies on well-matched large-scale datasets that are laborious in practice.
We introduce a novel noisy correspondence learning framework, namely textbfSelf-textbfReinforcing textbfErrors textbfMitigation (SREM)
arXiv Detail & Related papers (2023-12-27T09:03:43Z) - Improving Fair Training under Correlation Shifts [33.385118640843416]
In particular, when the bias between labels and sensitive groups changes, the fairness of the trained model is directly influenced and can worsen.
We analytically show that existing in-processing fair algorithms have fundamental limits in accuracy and group fairness.
We propose a novel pre-processing step that samples the input data to reduce correlation shifts.
arXiv Detail & Related papers (2023-02-05T07:23:35Z) - D-BIAS: A Causality-Based Human-in-the-Loop System for Tackling
Algorithmic Bias [57.87117733071416]
We propose D-BIAS, a visual interactive tool that embodies human-in-the-loop AI approach for auditing and mitigating social biases.
A user can detect the presence of bias against a group by identifying unfair causal relationships in the causal network.
For each interaction, say weakening/deleting a biased causal edge, the system uses a novel method to simulate a new (debiased) dataset.
arXiv Detail & Related papers (2022-08-10T03:41:48Z) - Understanding Unfairness in Fraud Detection through Model and Data Bias
Interactions [4.159343412286401]
We argue that algorithmic unfairness stems from interactions between models and biases in the data.
We study a set of hypotheses regarding the fairness-accuracy trade-offs that fairness-blind ML algorithms exhibit under different data bias settings.
arXiv Detail & Related papers (2022-07-13T15:18:30Z) - Sample Selection for Fair and Robust Training [28.94276265328868]
We propose a sample selection-based algorithm for fair and robust training.
We show that our algorithm obtains fairness and robustness better than or comparable to the state-of-the-art technique.
arXiv Detail & Related papers (2021-10-27T07:17:29Z) - Low-Regret Active learning [64.36270166907788]
We develop an online learning algorithm for identifying unlabeled data points that are most informative for training.
At the core of our work is an efficient algorithm for sleeping experts that is tailored to achieve low regret on predictable (easy) instances.
arXiv Detail & Related papers (2021-04-06T22:53:45Z) - Fairness Constraints in Semi-supervised Learning [56.48626493765908]
We develop a framework for fair semi-supervised learning, which is formulated as an optimization problem.
We theoretically analyze the source of discrimination in semi-supervised learning via bias, variance and noise decomposition.
Our method is able to achieve fair semi-supervised learning, and reach a better trade-off between accuracy and fairness than fair supervised learning.
arXiv Detail & Related papers (2020-09-14T04:25:59Z) - Causal Feature Selection for Algorithmic Fairness [61.767399505764736]
We consider fairness in the integration component of data management.
We propose an approach to identify a sub-collection of features that ensure the fairness of the dataset.
arXiv Detail & Related papers (2020-06-10T20:20:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.