Fairness through Optimization
- URL: http://arxiv.org/abs/2102.00311v2
- Date: Tue, 2 Feb 2021 01:55:39 GMT
- Title: Fairness through Optimization
- Authors: Violet Xinying Chen, J.N. Hooker
- Abstract summary: We argue that optimization models allow formulation of a wide range of fairness criteria as social welfare functions.
We show how optimization models can assist fairness-oriented decision making in the context of neural networks, support vector machines, and rule-based systems.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We propose optimization as a general paradigm for formalizing fairness in
AI-based decision models. We argue that optimization models allow formulation
of a wide range of fairness criteria as social welfare functions, while
enabling AI to take advantage of highly advanced solution technology. We show
how optimization models can assist fairness-oriented decision making in the
context of neural networks, support vector machines, and rule-based systems by
maximizing a social welfare function subject to appropriate constraints. In
particular, we state tractable optimization models for a variety of functions
that measure fairness or a combination of fairness and efficiency. These
include several inequality metrics, Rawlsian criteria, the McLoone and Hoover
indices, alpha fairness, the Nash and Kalai-Smorodinsky bargaining solutions,
combinations of Rawlsian and utilitarian criteria, and statistical bias
measures. All of these models can be efficiently solved by linear programming,
mixed integer/linear programming, or (in two cases) specialized convex
programming methods.
Related papers
- Learning Joint Models of Prediction and Optimization [56.04498536842065]
Predict-Then-Then framework uses machine learning models to predict unknown parameters of an optimization problem from features before solving.
This paper proposes an alternative method, in which optimal solutions are learned directly from the observable features by joint predictive models.
arXiv Detail & Related papers (2024-09-07T19:52:14Z) - Feature-Based Interpretable Surrogates for Optimization [0.8437187555622164]
In this work, we investigate how we can use more general optimization rules to increase interpretability.
The proposed rules do not map to a concrete solution but to a set of solutions characterized by common features.
In particular, we demonstrate the improvement in solution quality that our approach offers compared to existing interpretable surrogates for optimization.
arXiv Detail & Related papers (2024-09-03T13:12:49Z) - End-to-End Learning for Fair Multiobjective Optimization Under
Uncertainty [55.04219793298687]
The Predict-Then-Forecast (PtO) paradigm in machine learning aims to maximize downstream decision quality.
This paper extends the PtO methodology to optimization problems with nondifferentiable Ordered Weighted Averaging (OWA) objectives.
It shows how optimization of OWA functions can be effectively integrated with parametric prediction for fair and robust optimization under uncertainty.
arXiv Detail & Related papers (2024-02-12T16:33:35Z) - Predict-Then-Optimize by Proxy: Learning Joint Models of Prediction and
Optimization [59.386153202037086]
Predict-Then- framework uses machine learning models to predict unknown parameters of an optimization problem from features before solving.
This approach can be inefficient and requires handcrafted, problem-specific rules for backpropagation through the optimization step.
This paper proposes an alternative method, in which optimal solutions are learned directly from the observable features by predictive models.
arXiv Detail & Related papers (2023-11-22T01:32:06Z) - Global Optimization: A Machine Learning Approach [7.052596485478637]
Bertsimas and Ozturk (2023) proposed OCTHaGOn as a way of solving black-box global optimization problems.
We provide extensions to this approach by approximating the original problem using other MIO-representable ML models.
We show improvements in solution feasibility and optimality in the majority of instances.
arXiv Detail & Related papers (2023-11-03T06:33:38Z) - Stochastic Methods for AUC Optimization subject to AUC-based Fairness
Constraints [51.12047280149546]
A direct approach for obtaining a fair predictive model is to train the model through optimizing its prediction performance subject to fairness constraints.
We formulate the training problem of a fairness-aware machine learning model as an AUC optimization problem subject to a class of AUC-based fairness constraints.
We demonstrate the effectiveness of our approach on real-world data under different fairness metrics.
arXiv Detail & Related papers (2022-12-23T22:29:08Z) - Fast Feature Selection with Fairness Constraints [49.142308856826396]
We study the fundamental problem of selecting optimal features for model construction.
This problem is computationally challenging on large datasets, even with the use of greedy algorithm variants.
We extend the adaptive query model, recently proposed for the greedy forward selection for submodular functions, to the faster paradigm of Orthogonal Matching Pursuit for non-submodular functions.
The proposed algorithm achieves exponentially fast parallel run time in the adaptive query model, scaling much better than prior work.
arXiv Detail & Related papers (2022-02-28T12:26:47Z) - Approximate Bayesian Optimisation for Neural Networks [6.921210544516486]
A body of work has been done to automate machine learning algorithm to highlight the importance of model choice.
The necessity to solve the analytical tractability and the computational feasibility in a idealistic fashion enables to ensure the efficiency and the applicability.
arXiv Detail & Related papers (2021-08-27T19:03:32Z) - Modeling the Second Player in Distributionally Robust Optimization [90.25995710696425]
We argue for the use of neural generative models to characterize the worst-case distribution.
This approach poses a number of implementation and optimization challenges.
We find that the proposed approach yields models that are more robust than comparable baselines.
arXiv Detail & Related papers (2021-03-18T14:26:26Z) - Fair Bayesian Optimization [25.80374249896801]
We introduce a general constrained Bayesian optimization framework to optimize the performance of any machine learning (ML) model.
We apply BO with fairness constraints to a range of popular models, including random forests, boosting, and neural networks.
We show that our approach is competitive with specialized techniques that enforce model-specific fairness constraints.
arXiv Detail & Related papers (2020-06-09T08:31:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.