FairXGBoost: Fairness-aware Classification in XGBoost
- URL: http://arxiv.org/abs/2009.01442v2
- Date: Wed, 7 Oct 2020 05:14:38 GMT
- Title: FairXGBoost: Fairness-aware Classification in XGBoost
- Authors: Srinivasan Ravichandran, Drona Khurana, Bharath Venkatesh, Narayanan
Unny Edakunni
- Abstract summary: We propose a fair variant of XGBoost that enjoys all the advantages of XGBoost, while also matching the levels of fairness from bias-mitigation algorithms.
We provide an empirical analysis of our proposed method on standard benchmark datasets used in the fairness community.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Highly regulated domains such as finance have long favoured the use of
machine learning algorithms that are scalable, transparent, robust and yield
better performance. One of the most prominent examples of such an algorithm is
XGBoost. Meanwhile, there is also a growing interest in building fair and
unbiased models in these regulated domains and numerous bias-mitigation
algorithms have been proposed to this end. However, most of these
bias-mitigation methods are restricted to specific model families such as
logistic regression or support vector machine models, thus leaving modelers
with a difficult decision of choosing between fairness from the bias-mitigation
algorithms and scalability, transparency, performance from algorithms such as
XGBoost. We aim to leverage the best of both worlds by proposing a fair variant
of XGBoost that enjoys all the advantages of XGBoost, while also matching the
levels of fairness from the state-of-the-art bias-mitigation algorithms.
Furthermore, the proposed solution requires very little in terms of changes to
the original XGBoost library, thus making it easy for adoption. We provide an
empirical analysis of our proposed method on standard benchmark datasets used
in the fairness community.
Related papers
- Bi-discriminator Domain Adversarial Neural Networks with Class-Level
Gradient Alignment [87.8301166955305]
We propose a novel bi-discriminator domain adversarial neural network with class-level gradient alignment.
BACG resorts to gradient signals and second-order probability estimation for better alignment of domain distributions.
In addition, inspired by contrastive learning, we develop a memory bank-based variant, i.e. Fast-BACG, which can greatly shorten the training process.
arXiv Detail & Related papers (2023-10-21T09:53:17Z) - Boosting Fair Classifier Generalization through Adaptive Priority Reweighing [59.801444556074394]
A performance-promising fair algorithm with better generalizability is needed.
This paper proposes a novel adaptive reweighing method to eliminate the impact of the distribution shifts between training and test data on model generalizability.
arXiv Detail & Related papers (2023-09-15T13:04:55Z) - Fast Feature Selection with Fairness Constraints [49.142308856826396]
We study the fundamental problem of selecting optimal features for model construction.
This problem is computationally challenging on large datasets, even with the use of greedy algorithm variants.
We extend the adaptive query model, recently proposed for the greedy forward selection for submodular functions, to the faster paradigm of Orthogonal Matching Pursuit for non-submodular functions.
The proposed algorithm achieves exponentially fast parallel run time in the adaptive query model, scaling much better than prior work.
arXiv Detail & Related papers (2022-02-28T12:26:47Z) - FOLD-RM: A Scalable and Efficient Inductive Learning Algorithm for
Multi-Category Classification of Mixed Data [3.1981440103815717]
FOLD-RM is an automated inductive learning algorithm for learning default rules for mixed (numerical and categorical) data.
It generates an (explainable) answer set programming (ASP) rule set for multi-category classification tasks.
arXiv Detail & Related papers (2022-02-14T18:07:54Z) - Unpacking the Black Box: Regulating Algorithmic Decisions [1.283555556182245]
We propose a model of oversight over 'black-box' algorithms used in high-stakes applications such as lending, medical testing, or hiring.
We show that allowing for complex algorithms can improve welfare, but the gains depend on how the regulator regulates them.
arXiv Detail & Related papers (2021-10-05T23:20:25Z) - Individually Fair Gradient Boosting [86.1984206610373]
We consider the task of enforcing individual fairness in gradient boosting.
We show that our algorithm converges globally and generalizes.
We also demonstrate the efficacy of our algorithm on three ML problems susceptible to algorithmic bias.
arXiv Detail & Related papers (2021-03-31T03:06:57Z) - Towards Fair Knowledge Transfer for Imbalanced Domain Adaptation [61.317911756566126]
We propose a Towards Fair Knowledge Transfer framework to handle the fairness challenge in imbalanced cross-domain learning.
Specifically, a novel cross-domain mixup generation is exploited to augment the minority source set with target information to enhance fairness.
Our model significantly improves over 20% on two benchmarks in terms of the overall accuracy.
arXiv Detail & Related papers (2020-10-23T06:29:09Z) - Metrics and methods for a systematic comparison of fairness-aware
machine learning algorithms [0.0]
This study is the most comprehensive of its kind.
It considers fairness, predictive-performance, calibration quality, and speed of 28 different modelling pipelines.
We also found that fairness-aware algorithms can induce fairness without material drops in predictive power.
arXiv Detail & Related papers (2020-10-08T13:58:09Z) - Genetic programming approaches to learning fair classifiers [4.901632310846025]
We discuss current approaches to fairness and motivate proposals that incorporate fairness into genetic programming for classification.
The first is to incorporate a fairness objective into multi-objective optimization.
The second is to adapt lexicase selection to define cases dynamically over intersections of protected groups.
arXiv Detail & Related papers (2020-04-28T04:20:25Z) - On the Dual Formulation of Boosting Algorithms [92.74617630106559]
We show that the Lagrange problems of AdaBoost, LogitBoost and soft-marginBoost are all dual problems with generalized hinge loss entropy.
By looking at the dual problems of these boosting algorithms, we show that the success of boosting can be understood in terms of maintaining a better margin distribution.
arXiv Detail & Related papers (2009-01-23T02:14:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.