On Learning Fairness and Accuracy on Multiple Subgroups
- URL: http://arxiv.org/abs/2210.10837v2
- Date: Tue, 29 Nov 2022 15:20:07 GMT
- Title: On Learning Fairness and Accuracy on Multiple Subgroups
- Authors: Changjian Shui, Gezheng Xu, Qi Chen, Jiaqi Li, Charles Ling, Tal
Arbel, Boyu Wang, Christian Gagn\'e
- Abstract summary: We present a principled method for learning a fair predictor for all subgroups via formulating it as a bilevel objective.
Specifically, the subgroup specific predictors are learned in the lower-level through a small amount of data and the fair predictor.
In the upper-level, the fair predictor is updated to be close to all subgroup specific predictors.
- Score: 9.789933013990966
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: We propose an analysis in fair learning that preserves the utility of the
data while reducing prediction disparities under the criteria of group
sufficiency. We focus on the scenario where the data contains multiple or even
many subgroups, each with limited number of samples. As a result, we present a
principled method for learning a fair predictor for all subgroups via
formulating it as a bilevel objective. Specifically, the subgroup specific
predictors are learned in the lower-level through a small amount of data and
the fair predictor. In the upper-level, the fair predictor is updated to be
close to all subgroup specific predictors. We further prove that such a bilevel
objective can effectively control the group sufficiency and generalization
error. We evaluate the proposed framework on real-world datasets. Empirical
evidence suggests the consistently improved fair predictions, as well as the
comparable accuracy to the baselines.
Related papers
- How does promoting the minority fraction affect generalization? A theoretical study of the one-hidden-layer neural network on group imbalance [64.1656365676171]
Group imbalance has been a known problem in empirical risk minimization.
This paper quantifies the impact of individual groups on the sample complexity, the convergence rate, and the average and group-level testing performance.
arXiv Detail & Related papers (2024-03-12T04:38:05Z) - Fair Classification with Partial Feedback: An Exploration-Based Data Collection Approach [15.008626822593]
In many predictive contexts, true outcomes are only observed for samples that were positively classified in the past.
We present an approach that trains a classifier using available data and comes with a family of exploration strategies.
We show that this approach consistently boosts the quality of collected outcome data and improves the fraction of true positives for all groups.
arXiv Detail & Related papers (2024-02-17T17:09:19Z) - A structured regression approach for evaluating model performance across intersectional subgroups [53.91682617836498]
Disaggregated evaluation is a central task in AI fairness assessment, where the goal is to measure an AI system's performance across different subgroups.
We introduce a structured regression approach to disaggregated evaluation that we demonstrate can yield reliable system performance estimates even for very small subgroups.
arXiv Detail & Related papers (2024-01-26T14:21:45Z) - Bias Amplification Enhances Minority Group Performance [10.380812738348899]
We propose BAM, a novel two-stage training algorithm.
In the first stage, the model is trained using a bias amplification scheme via introducing a learnable auxiliary variable for each training sample.
In the second stage, we upweight the samples that the bias-amplified model misclassifies, and then continue training the same model on the reweighted dataset.
arXiv Detail & Related papers (2023-09-13T04:40:08Z) - Fairness Aware Counterfactuals for Subgroups [8.593488857185678]
We present Fairness Aware Counterfactuals for Subgroups (FACTS), a framework for auditing subgroup fairness through counterfactual explanations.
We aim to formulate different aspects of the difficulty of individuals in certain subgroups to achieve recourse.
We introduce notions of subgroup fairness that are robust, if not totally oblivious, to the cost of achieving recourse.
arXiv Detail & Related papers (2023-06-26T18:03:56Z) - Leveraging Structure for Improved Classification of Grouped Biased Data [8.121462458089143]
We consider semi-supervised binary classification for applications in which data points are naturally grouped.
We derive a semi-supervised algorithm that explicitly leverages the structure to learn an optimal, group-aware, probability-outputd classifier.
arXiv Detail & Related papers (2022-12-07T15:18:21Z) - Addressing Missing Sources with Adversarial Support-Matching [8.53946780558779]
We investigate a scenario in which the absence of certain data is linked to the second level of a two-level hierarchy in the data.
Inspired by the idea of protected groups from algorithmic fairness, we refer to the partitions carved by this second level as "subgroups"
We make use of an additional, diverse but unlabeled dataset, called the "deployment set", to learn a representation that is invariant to subgroup.
arXiv Detail & Related papers (2022-03-24T16:19:19Z) - Fair Group-Shared Representations with Normalizing Flows [68.29997072804537]
We develop a fair representation learning algorithm which is able to map individuals belonging to different groups in a single group.
We show experimentally that our methodology is competitive with other fair representation learning algorithms.
arXiv Detail & Related papers (2022-01-17T10:49:49Z) - Towards Group Robustness in the presence of Partial Group Labels [61.33713547766866]
spurious correlations between input samples and the target labels wrongly direct the neural network predictions.
We propose an algorithm that optimize for the worst-off group assignments from a constraint set.
We show improvements in the minority group's performance while preserving overall aggregate accuracy across groups.
arXiv Detail & Related papers (2022-01-10T22:04:48Z) - Representation Matters: Assessing the Importance of Subgroup Allocations
in Training Data [85.43008636875345]
We show that diverse representation in training data is key to increasing subgroup performances and achieving population level objectives.
Our analysis and experiments describe how dataset compositions influence performance and provide constructive results for using trends in existing data, alongside domain knowledge, to help guide intentional, objective-aware dataset design.
arXiv Detail & Related papers (2021-03-05T00:27:08Z) - Towards Model-Agnostic Post-Hoc Adjustment for Balancing Ranking
Fairness and Algorithm Utility [54.179859639868646]
Bipartite ranking aims to learn a scoring function that ranks positive individuals higher than negative ones from labeled data.
There have been rising concerns on whether the learned scoring function can cause systematic disparity across different protected groups.
We propose a model post-processing framework for balancing them in the bipartite ranking scenario.
arXiv Detail & Related papers (2020-06-15T10:08:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.