Fairness Hub Technical Briefs: AUC Gap
- URL: http://arxiv.org/abs/2309.12371v2
- Date: Mon, 25 Sep 2023 21:04:44 GMT
- Title: Fairness Hub Technical Briefs: AUC Gap
- Authors: Jinsook Lee, Chris Brooks, Renzhe Yu, Rene Kizilcec
- Abstract summary: To measure bias, we encourage teams to consider using AUC Gap: the absolute difference between the highest and lowest test AUC for subgroups.
It is agnostic to the AI/ML algorithm used and it captures the disparity in model performance for any number of subgroups.
The teams use a wide range of AI/ML models in pursuit of a common goal of doubling math achievement in low-income middle schools.
- Score: 0.6827423171182154
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: To measure bias, we encourage teams to consider using AUC Gap: the absolute
difference between the highest and lowest test AUC for subgroups (e.g., gender,
race, SES, prior knowledge). It is agnostic to the AI/ML algorithm used and it
captures the disparity in model performance for any number of subgroups, which
enables non-binary fairness assessments such as for intersectional identity
groups. The teams use a wide range of AI/ML models in pursuit of a common goal
of doubling math achievement in low-income middle schools. Ensuring that the
models, which are trained on datasets collected in many different contexts, do
not introduce or amplify biases is important for achieving the goal. We offer
here a versatile and easy-to-compute measure of model bias for all the teams in
order to create a common benchmark and an analytical basis for sharing what
strategies have worked for different teams.
Related papers
- Semi-Supervised One-Shot Imitation Learning [83.94646047695412]
One-shot Imitation Learning aims to imbue AI agents with the ability to learn a new task from a single demonstration.
We introduce the semi-supervised OSIL problem setting, where the learning agent is presented with a large dataset of trajectories.
We develop an algorithm specifically applicable to this semi-supervised OSIL setting.
arXiv Detail & Related papers (2024-08-09T18:11:26Z) - Improving Generalization of Alignment with Human Preferences through
Group Invariant Learning [56.19242260613749]
Reinforcement Learning from Human Feedback (RLHF) enables the generation of responses more aligned with human preferences.
Previous work shows that Reinforcement Learning (RL) often exploits shortcuts to attain high rewards and overlooks challenging samples.
We propose a novel approach that can learn a consistent policy via RL across various data groups or domains.
arXiv Detail & Related papers (2023-10-18T13:54:15Z) - Universal Semi-supervised Model Adaptation via Collaborative Consistency
Training [92.52892510093037]
We introduce a realistic and challenging domain adaptation problem called Universal Semi-supervised Model Adaptation (USMA)
We propose a collaborative consistency training framework that regularizes the prediction consistency between two models.
Experimental results demonstrate the effectiveness of our method on several benchmark datasets.
arXiv Detail & Related papers (2023-07-07T08:19:40Z) - Change is Hard: A Closer Look at Subpopulation Shift [48.0369745740936]
We propose a unified framework that dissects and explains common shifts in subgroups.
We then establish a benchmark of 20 state-of-the-art algorithms evaluated on 12 real-world datasets in vision, language, and healthcare domains.
arXiv Detail & Related papers (2023-02-23T18:59:56Z) - Outlier-Robust Group Inference via Gradient Space Clustering [50.87474101594732]
Existing methods can improve the worst-group performance, but they require group annotations, which are often expensive and sometimes infeasible to obtain.
We address the problem of learning group annotations in the presence of outliers by clustering the data in the space of gradients of the model parameters.
We show that data in the gradient space has a simpler structure while preserving information about minority groups and outliers, making it suitable for standard clustering methods like DBSCAN.
arXiv Detail & Related papers (2022-10-13T06:04:43Z) - Unifying Language Learning Paradigms [96.35981503087567]
We present a unified framework for pre-training models that are universally effective across datasets and setups.
We show how different pre-training objectives can be cast as one another and how interpolating between different objectives can be effective.
Our model also achieve strong results at in-context learning, outperforming 175B GPT-3 on zero-shot SuperGLUE and tripling the performance of T5-XXL on one-shot summarization.
arXiv Detail & Related papers (2022-05-10T19:32:20Z) - Fair Group-Shared Representations with Normalizing Flows [68.29997072804537]
We develop a fair representation learning algorithm which is able to map individuals belonging to different groups in a single group.
We show experimentally that our methodology is competitive with other fair representation learning algorithms.
arXiv Detail & Related papers (2022-01-17T10:49:49Z) - Fairness for AUC via Feature Augmentation [25.819342066717002]
We study fairness in the context of classification where the performance is measured by the area under the curve (AUC) of the receiver operating characteristic.
We develop a novel approach, fairAUC, based on feature augmentation (adding features) to mitigate bias between identifiable groups.
arXiv Detail & Related papers (2021-11-24T22:32:19Z) - Unified Group Fairness on Federated Learning [22.143427873780404]
Federated learning (FL) has emerged as an important machine learning paradigm where a global model is trained based on private data from distributed clients.
Recent researches focus on achieving fairness among clients, but they ignore the fairness towards different groups formed by sensitive attribute(s) (e.g., gender and/or race)
We propose a novel FL algorithm, named Group Distributionally Robust Federated Averaging (G-DRFA), which mitigates the distribution shift across groups with theoretical analysis of convergence rate.
arXiv Detail & Related papers (2021-11-09T08:21:38Z) - Few-shot Learning with Global Relatedness Decoupled-Distillation [47.78903405454224]
We propose a new Global Relatedness Decoupled-Distillation (GRDD) method using the global category knowledge and the Relatedness Decoupled-Distillation (RDD) strategy.
Our GRDD learns new visual concepts quickly by imitating the habit of humans, i.e. learning from the deep knowledge distilled from the teacher.
arXiv Detail & Related papers (2021-07-12T17:01:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.