FAIR-Ensemble: When Fairness Naturally Emerges From Deep Ensembling
- URL: http://arxiv.org/abs/2303.00586v2
- Date: Wed, 20 Dec 2023 22:54:48 GMT
- Title: FAIR-Ensemble: When Fairness Naturally Emerges From Deep Ensembling
- Authors: Wei-Yin Ko, Daniel D'souza, Karina Nguyen, Randall Balestriero, Sara
Hooker
- Abstract summary: Ensembling multiple Deep Neural Networks (DNNs) is a simple and effective way to improve top-line metrics.
In this work, we explore the impact of ensembling on subgroup performances.
- Score: 17.731480052857158
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Ensembling multiple Deep Neural Networks (DNNs) is a simple and effective way
to improve top-line metrics and to outperform a larger single model. In this
work, we go beyond top-line metrics and instead explore the impact of
ensembling on subgroup performances. Surprisingly, we observe that even with a
simple homogeneous ensemble -- all the individual DNNs share the same training
set, architecture, and design choices -- the minority group performance
disproportionately improves with the number of models compared to the majority
group, i.e. fairness naturally emerges from ensembling. Even more surprising,
we find that this gain keeps occurring even when a large number of models is
considered, e.g. $20$, despite the fact that the average performance of the
ensemble plateaus with fewer models. Our work establishes that simple DNN
ensembles can be a powerful tool for alleviating disparate impact from DNN
classifiers, thus curbing algorithmic harm. We also explore why this is the
case. We find that even in homogeneous ensembles, varying the sources of
stochasticity through parameter initialization, mini-batch sampling, and
data-augmentation realizations, results in different fairness outcomes.
Related papers
- Rethinking Fair Graph Neural Networks from Re-balancing [26.70771023446706]
We find that simple re-balancing methods can easily match or surpass existing fair GNN methods.
We propose FairGB, Fair Graph Neural Network via re-Balancing, which mitigates the unfairness of GNNs by group balancing.
arXiv Detail & Related papers (2024-07-16T11:39:27Z) - Enhancing Few-shot NER with Prompt Ordering based Data Augmentation [59.69108119752584]
We propose a Prompt Ordering based Data Augmentation (PODA) method to improve the training of unified autoregressive generation frameworks.
Experimental results on three public NER datasets and further analyses demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2023-05-19T16:25:43Z) - Joint Training of Deep Ensembles Fails Due to Learner Collusion [61.557412796012535]
Ensembles of machine learning models have been well established as a powerful method of improving performance over a single model.
Traditionally, ensembling algorithms train their base learners independently or sequentially with the goal of optimizing their joint performance.
We show that directly minimizing the loss of the ensemble appears to rarely be applied in practice.
arXiv Detail & Related papers (2023-01-26T18:58:07Z) - Deep Negative Correlation Classification [82.45045814842595]
Existing deep ensemble methods naively train many different models and then aggregate their predictions.
We propose deep negative correlation classification (DNCC)
DNCC yields a deep classification ensemble where the individual estimator is both accurate and negatively correlated.
arXiv Detail & Related papers (2022-12-14T07:35:20Z) - Simplicity Bias Leads to Amplified Performance Disparities [8.60453031364566]
We show that SGD-trained models have a bias towards simplicity, leading them to prioritize learning a majority class.
A model may prioritize any class or group of the dataset that it finds simple-at the expense of what it finds complex.
arXiv Detail & Related papers (2022-12-13T15:24:41Z) - How to train your draGAN: A task oriented solution to imbalanced
classification [15.893327571516016]
This paper proposes a unique, performance-oriented, data-generating strategy that utilizes a new architecture, coined draGAN.
The samples are generated with the objective of optimizing the classification model's performance, rather than similarity to the real data.
Empirically we show the superiority of draGAN, but also highlight some of its shortcomings.
arXiv Detail & Related papers (2022-11-18T07:37:34Z) - Deep Ensembles Work, But Are They Necessary? [19.615082441403946]
Ensembling neural networks is an effective way to increase accuracy.
Recent work suggests that deep ensembles may offer benefits beyond predictive power.
We show that a single (but larger) neural network can replicate these qualities.
arXiv Detail & Related papers (2022-02-14T19:01:01Z) - Sparse MoEs meet Efficient Ensembles [49.313497379189315]
We study the interplay of two popular classes of such models: ensembles of neural networks and sparse mixture of experts (sparse MoEs)
We present Efficient Ensemble of Experts (E$3$), a scalable and simple ensemble of sparse MoEs that takes the best of both classes of models, while using up to 45% fewer FLOPs than a deep ensemble.
arXiv Detail & Related papers (2021-10-07T11:58:35Z) - Examining and Combating Spurious Features under Distribution Shift [94.31956965507085]
We define and analyze robust and spurious representations using the information-theoretic concept of minimal sufficient statistics.
We prove that even when there is only bias of the input distribution, models can still pick up spurious features from their training data.
Inspired by our analysis, we demonstrate that group DRO can fail when groups do not directly account for various spurious correlations.
arXiv Detail & Related papers (2021-06-14T05:39:09Z) - When Ensembling Smaller Models is More Efficient than Single Large
Models [52.38997176317532]
We show that ensembles can outperform single models with both higher accuracy and requiring fewer total FLOPs to compute.
This presents an interesting observation that output diversity in ensembling can often be more efficient than training larger models.
arXiv Detail & Related papers (2020-05-01T18:56:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.