Fairness-guided SMT-based Rectification of Decision Trees and Random
Forests
- URL: http://arxiv.org/abs/2011.11001v1
- Date: Sun, 22 Nov 2020 12:30:27 GMT
- Title: Fairness-guided SMT-based Rectification of Decision Trees and Random
Forests
- Authors: Jiang Zhang, Ivan Beschastnikh, Sergey Mechtaev, Abhik Roychoudhury
- Abstract summary: Our approach converts any decision tree or random forest into a fair one with respect to a specific data set, fairness criteria, and sensitive attributes.
Our experiments on the well-known adult dataset from UC Irvine demonstrate that FairRepair scales to realistic decision trees and random forests.
Since our fairness-guided repair technique repairs decision trees and random forests obtained from a given (unfair) data-set, it can help to identify and rectify biases in decision-making in an organisation.
- Score: 14.423550468823152
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Data-driven decision making is gaining prominence with the popularity of
various machine learning models. Unfortunately, real-life data used in machine
learning training may capture human biases, and as a result the learned models
may lead to unfair decision making. In this paper, we provide a solution to
this problem for decision trees and random forests. Our approach converts any
decision tree or random forest into a fair one with respect to a specific data
set, fairness criteria, and sensitive attributes. The \emph{FairRepair} tool,
built based on our approach, is inspired by automated program repair techniques
for traditional programs. It uses an SMT solver to decide which paths in the
decision tree could have their outcomes flipped to improve the fairness of the
model. Our experiments on the well-known adult dataset from UC Irvine
demonstrate that FairRepair scales to realistic decision trees and random
forests. Furthermore, FairRepair provides formal guarantees about soundness and
completeness of finding a repair. Since our fairness-guided repair technique
repairs decision trees and random forests obtained from a given (unfair)
data-set, it can help to identify and rectify biases in decision-making in an
organisation.
Related papers
- SKADA-Bench: Benchmarking Unsupervised Domain Adaptation Methods with Realistic Validation [55.87169702896249]
Unsupervised Domain Adaptation (DA) consists of adapting a model trained on a labeled source domain to perform well on an unlabeled target domain with some data distribution shift.
We propose a framework to evaluate DA methods and present a fair evaluation of existing shallow algorithms, including reweighting, mapping, and subspace alignment.
Our benchmark highlights the importance of realistic validation and provides practical guidance for real-life applications.
arXiv Detail & Related papers (2024-07-16T12:52:29Z) - An Interpretable Client Decision Tree Aggregation process for Federated Learning [7.8973037023478785]
We propose an Interpretable Client Decision Tree aggregation process for Federated Learning scenarios.
This model is based on aggregating multiple decision paths of the decision trees and can be used on different decision tree types, such as ID3 and CART.
We carry out the experiments within four datasets, and the analysis shows that the tree built with the model improves the local models, and outperforms the state-of-the-art.
arXiv Detail & Related papers (2024-04-03T06:53:56Z) - Scaling Laws Beyond Backpropagation [64.0476282000118]
We study the ability of Direct Feedback Alignment to train causal decoder-only Transformers efficiently.
We find that DFA fails to offer more efficient scaling than backpropagation.
arXiv Detail & Related papers (2022-10-26T10:09:14Z) - Optimal Decision Diagrams for Classification [68.72078059880018]
We study the training of optimal decision diagrams from a mathematical programming perspective.
We introduce a novel mixed-integer linear programming model for training.
We show how this model can be easily extended for fairness, parsimony, and stability notions.
arXiv Detail & Related papers (2022-05-28T18:31:23Z) - EiFFFeL: Enforcing Fairness in Forests by Flipping Leaves [5.08078625937586]
We propose a fairness enforcing approach called EiFFFeL:Enforcing Fairness in Forests by Flipping Leaves.
Experimental results show that our approach achieves a user defined group fairness degree without losing a significant amount of accuracy.
arXiv Detail & Related papers (2021-12-29T07:48:38Z) - Leveraging Expert Consistency to Improve Algorithmic Decision Support [62.61153549123407]
We explore the use of historical expert decisions as a rich source of information that can be combined with observed outcomes to narrow the construct gap.
We propose an influence function-based methodology to estimate expert consistency indirectly when each case in the data is assessed by a single expert.
Our empirical evaluation, using simulations in a clinical setting and real-world data from the child welfare domain, indicates that the proposed approach successfully narrows the construct gap.
arXiv Detail & Related papers (2021-01-24T05:40:29Z) - Fair Training of Decision Tree Classifiers [6.381149074212897]
We study the problem of formally verifying individual fairness of decision tree ensembles.
In our approach, fairness verification and fairness-aware training both rely on a notion of stability of a classification model.
arXiv Detail & Related papers (2021-01-04T12:04:22Z) - Beyond Individual and Group Fairness [90.4666341812857]
We present a new data-driven model of fairness that is guided by the unfairness complaints received by the system.
Our model supports multiple fairness criteria and takes into account their potential incompatibilities.
arXiv Detail & Related papers (2020-08-21T14:14:44Z) - Rectified Decision Trees: Exploring the Landscape of Interpretable and
Effective Machine Learning [66.01622034708319]
We propose a knowledge distillation based decision trees extension, dubbed rectified decision trees (ReDT)
We extend the splitting criteria and the ending condition of the standard decision trees, which allows training with soft labels.
We then train the ReDT based on the soft label distilled from a well-trained teacher model through a novel jackknife-based method.
arXiv Detail & Related papers (2020-08-21T10:45:25Z) - Learning Representations for Axis-Aligned Decision Forests through Input
Perturbation [2.755007887718791]
Axis-aligned decision forests have long been the leading class of machine learning algorithms.
Despite their widespread use and rich history, decision forests to date fail to consume raw structured data.
We present a novel but intuitive proposal to achieve representation learning for decision forests.
arXiv Detail & Related papers (2020-07-29T11:56:38Z) - How fair can we go in machine learning? Assessing the boundaries of
fairness in decision trees [0.12891210250935145]
We present the first methodology that allows to explore the statistical limits of bias mitigation interventions.
We focus our study on decision tree classifiers since they are widely accepted in machine learning.
We conclude experimentally that our method can optimize decision tree models by being fairer with a small cost of the classification error.
arXiv Detail & Related papers (2020-06-22T16:28:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.