Fairness in Multi-Task Learning via Wasserstein Barycenters
- URL: http://arxiv.org/abs/2306.10155v2
- Date: Thu, 6 Jul 2023 09:37:36 GMT
- Title: Fairness in Multi-Task Learning via Wasserstein Barycenters
- Authors: Fran\c{c}ois Hu, Philipp Ratz, Arthur Charpentier
- Abstract summary: Algorithmic Fairness is an established field in machine learning that aims to reduce biases in data.
We develop a method that extends the definition of Strong Demographic Parity to multi-task learning using multi-marginal Wasserstein barycenters.
Our approach provides a closed form solution for the optimal fair multi-task predictor including both regression and binary classification tasks.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Algorithmic Fairness is an established field in machine learning that aims to
reduce biases in data. Recent advances have proposed various methods to ensure
fairness in a univariate environment, where the goal is to de-bias a single
task. However, extending fairness to a multi-task setting, where more than one
objective is optimised using a shared representation, remains underexplored. To
bridge this gap, we develop a method that extends the definition of Strong
Demographic Parity to multi-task learning using multi-marginal Wasserstein
barycenters. Our approach provides a closed form solution for the optimal fair
multi-task predictor including both regression and binary classification tasks.
We develop a data-driven estimation procedure for the solution and run
numerical experiments on both synthetic and real datasets. The empirical
results highlight the practical value of our post-processing methodology in
promoting fair decision-making.
Related papers
- Multi-Output Distributional Fairness via Post-Processing [47.94071156898198]
We introduce a post-processing method for multi-output models to enhance a model's distributional parity, a task-agnostic fairness measure.
Our method employs an optimal transport mapping to move a model's outputs across different groups towards their empirical Wasserstein barycenter.
arXiv Detail & Related papers (2024-08-31T22:41:26Z) - Estimating Barycenters of Distributions with Neural Optimal Transport [93.28746685008093]
We propose a new scalable approach for solving the Wasserstein barycenter problem.
Our methodology is based on the recent Neural OT solver.
We also establish theoretical error bounds for our proposed approach.
arXiv Detail & Related papers (2024-02-06T09:17:07Z) - Multi-task Bias-Variance Trade-off Through Functional Constraints [102.64082402388192]
Multi-task learning aims to acquire a set of functions that perform well for diverse tasks.
In this paper we draw intuition from the two extreme learning scenarios -- a single function for all tasks, and a task-specific function that ignores the other tasks.
We introduce a constrained learning formulation that enforces domain specific solutions to a central function.
arXiv Detail & Related papers (2022-10-27T16:06:47Z) - Mitigating Gradient Bias in Multi-objective Learning: A Provably Convergent Stochastic Approach [38.76462300149459]
We develop a Multi-objective Correction (MoCo) method for multi-objective gradient optimization.
The unique feature of our method is that it can guarantee convergence without increasing the non fairness gradient.
arXiv Detail & Related papers (2022-10-23T05:54:26Z) - Uncertainty-Aware Meta-Learning for Multimodal Task Distributions [3.7470451129384825]
We present UnLiMiTD (uncertainty-aware meta-learning for multimodal task distributions)
We take a probabilistic perspective and train a parametric, tuneable distribution over tasks on the meta-dataset.
We demonstrate that UnLiMiTD's predictions compare favorably to, and outperform in most cases, the standard baselines.
arXiv Detail & Related papers (2022-10-04T20:02:25Z) - On Modality Bias Recognition and Reduction [70.69194431713825]
We study the modality bias problem in the context of multi-modal classification.
We propose a plug-and-play loss function method, whereby the feature space for each label is adaptively learned.
Our method yields remarkable performance improvements compared with the baselines.
arXiv Detail & Related papers (2022-02-25T13:47:09Z) - Conditional Meta-Learning of Linear Representations [57.90025697492041]
Standard meta-learning for representation learning aims to find a common representation to be shared across multiple tasks.
In this work we overcome this issue by inferring a conditioning function, mapping the tasks' side information into a representation tailored to the task at hand.
We propose a meta-algorithm capable of leveraging this advantage in practice.
arXiv Detail & Related papers (2021-03-30T12:02:14Z) - Multi-task Learning by Leveraging the Semantic Information [14.397128692867799]
We propose to leverage the label information in multi-task learning by exploring the semantic conditional relations among tasks.
Our analysis also leads to a concrete algorithm that jointly matches the semantic distribution and controls label distribution divergence.
arXiv Detail & Related papers (2021-03-03T17:36:35Z) - High-Dimensional Multi-Task Averaging and Application to Kernel Mean
Embedding [0.0]
We propose an improved estimator for the multi-task averaging problem.
We prove theoretically that this approach provides a reduction in mean squared error.
An application of this approach is the estimation of multiple kernel mean embeddings.
arXiv Detail & Related papers (2020-11-13T07:31:30Z) - Towards Model-Agnostic Post-Hoc Adjustment for Balancing Ranking
Fairness and Algorithm Utility [54.179859639868646]
Bipartite ranking aims to learn a scoring function that ranks positive individuals higher than negative ones from labeled data.
There have been rising concerns on whether the learned scoring function can cause systematic disparity across different protected groups.
We propose a model post-processing framework for balancing them in the bipartite ranking scenario.
arXiv Detail & Related papers (2020-06-15T10:08:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.