iFlipper: Label Flipping for Individual Fairness
- URL: http://arxiv.org/abs/2209.07047v1
- Date: Thu, 15 Sep 2022 05:02:01 GMT
- Title: iFlipper: Label Flipping for Individual Fairness
- Authors: Hantian Zhang, Ki Hyun Tae, Jaeyoung Park, Xu Chu, Steven Euijong
Whang
- Abstract summary: We show that label flipping is an effective pre-processing technique for improving individual fairness.
We propose an approximate linear programming algorithm and provide theoretical guarantees on how close its result is to the optimal solution.
Experiments on real datasets show that iFlipper significantly outperforms other pre-processing baselines in terms of individual fairness.
- Score: 16.50058737985628
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: As machine learning becomes prevalent, mitigating any unfairness present in
the training data becomes critical. Among the various notions of fairness, this
paper focuses on the well-known individual fairness, which states that similar
individuals should be treated similarly. While individual fairness can be
improved when training a model (in-processing), we contend that fixing the data
before model training (pre-processing) is a more fundamental solution. In
particular, we show that label flipping is an effective pre-processing
technique for improving individual fairness. Our system iFlipper solves the
optimization problem of minimally flipping labels given a limit to the
individual fairness violations, where a violation occurs when two similar
examples in the training data have different labels. We first prove that the
problem is NP-hard. We then propose an approximate linear programming algorithm
and provide theoretical guarantees on how close its result is to the optimal
solution in terms of the number of label flips. We also propose techniques for
making the linear programming solution more optimal without exceeding the
violations limit. Experiments on real datasets show that iFlipper significantly
outperforms other pre-processing baselines in terms of individual fairness and
accuracy on unseen test sets. In addition, iFlipper can be combined with
in-processing techniques for even better results.
Related papers
- Improving Fair Training under Correlation Shifts [33.385118640843416]
In particular, when the bias between labels and sensitive groups changes, the fairness of the trained model is directly influenced and can worsen.
We analytically show that existing in-processing fair algorithms have fundamental limits in accuracy and group fairness.
We propose a novel pre-processing step that samples the input data to reduce correlation shifts.
arXiv Detail & Related papers (2023-02-05T07:23:35Z) - SoftMatch: Addressing the Quantity-Quality Trade-off in Semi-supervised
Learning [101.86916775218403]
This paper revisits the popular pseudo-labeling methods via a unified sample weighting formulation.
We propose SoftMatch to overcome the trade-off by maintaining both high quantity and high quality of pseudo-labels during training.
In experiments, SoftMatch shows substantial improvements across a wide variety of benchmarks, including image, text, and imbalanced classification.
arXiv Detail & Related papers (2023-01-26T03:53:25Z) - Communication-Efficient Robust Federated Learning with Noisy Labels [144.31995882209932]
Federated learning (FL) is a promising privacy-preserving machine learning paradigm over distributed located data.
We propose a learning-based reweighting approach to mitigate the effect of noisy labels in FL.
Our approach has shown superior performance on several real-world datasets compared to various baselines.
arXiv Detail & Related papers (2022-06-11T16:21:17Z) - Normalise for Fairness: A Simple Normalisation Technique for Fairness in Regression Machine Learning Problems [46.93320580613236]
We present a simple, yet effective method based on normalisation (FaiReg) for regression problems.
We compare it with two standard methods for fairness, namely data balancing and adversarial training.
The results show the superior performance of diminishing the effects of unfairness better than data balancing.
arXiv Detail & Related papers (2022-02-02T12:26:25Z) - A Pre-processing Method for Fairness in Ranking [0.0]
We propose a fair ranking framework that evaluates the order of training data in a pairwise manner.
We show that our method outperforms the existing methods in the trade-off between accuracy and fairness over real-world datasets.
arXiv Detail & Related papers (2021-10-29T02:55:32Z) - Post-processing for Individual Fairness [23.570995756189266]
Post-processing in algorithmic fairness is a versatile approach for correcting bias in ML systems that are already used in production.
We consider a setting where the learner only has access to the predictions of the original model and a similarity graph between individuals, guiding the desired fairness constraints.
Our algorithms correct individual biases in large-scale NLP models such as BERT, while preserving accuracy.
arXiv Detail & Related papers (2021-10-26T15:51:48Z) - BiFair: Training Fair Models with Bilevel Optimization [8.2509884277533]
We develop a new training algorithm, named BiFair, which jointly minimizes for a utility, and a fairness loss of interest.
Our algorithm consistently performs better, i.e., we reach to better values of a given fairness metric under same, or higher accuracy.
arXiv Detail & Related papers (2021-06-03T22:36:17Z) - Can Active Learning Preemptively Mitigate Fairness Issues? [66.84854430781097]
dataset bias is one of the prevailing causes of unfairness in machine learning.
We study whether models trained with uncertainty-based ALs are fairer in their decisions with respect to a protected class.
We also explore the interaction of algorithmic fairness methods such as gradient reversal (GRAD) and BALD.
arXiv Detail & Related papers (2021-04-14T14:20:22Z) - Fairness in Semi-supervised Learning: Unlabeled Data Help to Reduce
Discrimination [53.3082498402884]
A growing specter in the rise of machine learning is whether the decisions made by machine learning models are fair.
We present a framework of fair semi-supervised learning in the pre-processing phase, including pseudo labeling to predict labels for unlabeled data.
A theoretical decomposition analysis of bias, variance and noise highlights the different sources of discrimination and the impact they have on fairness in semi-supervised learning.
arXiv Detail & Related papers (2020-09-25T05:48:56Z) - Fairness Constraints in Semi-supervised Learning [56.48626493765908]
We develop a framework for fair semi-supervised learning, which is formulated as an optimization problem.
We theoretically analyze the source of discrimination in semi-supervised learning via bias, variance and noise decomposition.
Our method is able to achieve fair semi-supervised learning, and reach a better trade-off between accuracy and fairness than fair supervised learning.
arXiv Detail & Related papers (2020-09-14T04:25:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.