Learning to Split for Automatic Bias Detection
- URL: http://arxiv.org/abs/2204.13749v1
- Date: Thu, 28 Apr 2022 19:41:08 GMT
- Title: Learning to Split for Automatic Bias Detection
- Authors: Yujia Bao, Regina Barzilay
- Abstract summary: Learning to Split (ls) is an algorithm for automatic bias detection.
We evaluate our approach on Beer Review, CelebA and MNLI.
- Score: 39.353850990332525
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Classifiers are biased when trained on biased datasets. As a remedy, we
propose Learning to Split (ls), an algorithm for automatic bias detection.
Given a dataset with input-label pairs, ls learns to split this dataset so that
predictors trained on the training split generalize poorly to the testing
split. This performance gap provides a proxy for measuring the degree of bias
in the learned features and can therefore be used to reduce biases. Identifying
non-generalizable splits is challenging as we don't have any explicit
annotations about how to split. In this work, we show that the prediction
correctness of the testing example can be used as a source of weak supervision:
generalization performance will drop if we move examples that are predicted
correctly away from the testing split, leaving only those that are
mispredicted. We evaluate our approach on Beer Review, Waterbirds, CelebA and
MNLI. Empirical results show that ls is able to generate astonishingly
challenging splits that correlate with human-identified biases. Moreover, we
demonstrate that combining robust learning algorithms (such as group DRO) with
splits identified by ls enables automatic de-biasing. Compared with previous
state-of-the-arts, we substantially improves the worst-group performance (23.4%
on average) when the source of biases is unknown during training and
validation.
Related papers
- Distributional bias compromises leave-one-out cross-validation [0.6656737591902598]
Cross-validation is a common method for estimating the predictive performance of machine learning models.
We show that an approach called "leave-one-out cross-validation" creates a negative correlation between the average label of each training fold and the label of its corresponding test instance.
We propose a generalizable rebalanced cross-validation approach that corrects for distributional bias.
arXiv Detail & Related papers (2024-06-03T15:47:34Z) - Revisiting the Dataset Bias Problem from a Statistical Perspective [72.94990819287551]
We study the "dataset bias" problem from a statistical standpoint.
We identify the main cause of the problem as the strong correlation between a class attribute u and a non-class attribute b.
We propose to mitigate dataset bias via either weighting the objective of each sample n by frac1p(u_n|b_n) or sampling that sample with a weight proportional to frac1p(u_n|b_n).
arXiv Detail & Related papers (2024-02-05T22:58:06Z) - Noisy Correspondence Learning with Self-Reinforcing Errors Mitigation [63.180725016463974]
Cross-modal retrieval relies on well-matched large-scale datasets that are laborious in practice.
We introduce a novel noisy correspondence learning framework, namely textbfSelf-textbfReinforcing textbfErrors textbfMitigation (SREM)
arXiv Detail & Related papers (2023-12-27T09:03:43Z) - Correcting Underrepresentation and Intersectional Bias for Classification [49.1574468325115]
We consider the problem of learning from data corrupted by underrepresentation bias.
We show that with a small amount of unbiased data, we can efficiently estimate the group-wise drop-out rates.
We show that our algorithm permits efficient learning for model classes of finite VC dimension.
arXiv Detail & Related papers (2023-06-19T18:25:44Z) - Beyond Random Split for Assessing Statistical Model Performance [0.0]
We analyze strategies based on predictors' variability to split in training and testing sets.
Such strategies aim at guaranteeing the inclusion of rare or unusual examples with a minimal loss of the population's representativeness.
Preliminary results showed the importance of applying the three alternative strategies to the Monte Carlo splitting strategy.
arXiv Detail & Related papers (2022-09-04T22:24:35Z) - Relieving Long-tailed Instance Segmentation via Pairwise Class Balance [85.53585498649252]
Long-tailed instance segmentation is a challenging task due to the extreme imbalance of training samples among classes.
It causes severe biases of the head classes (with majority samples) against the tailed ones.
We propose a novel Pairwise Class Balance (PCB) method, built upon a confusion matrix which is updated during training to accumulate the ongoing prediction preferences.
arXiv Detail & Related papers (2022-01-08T07:48:36Z) - Examining and Combating Spurious Features under Distribution Shift [94.31956965507085]
We define and analyze robust and spurious representations using the information-theoretic concept of minimal sufficient statistics.
We prove that even when there is only bias of the input distribution, models can still pick up spurious features from their training data.
Inspired by our analysis, we demonstrate that group DRO can fail when groups do not directly account for various spurious correlations.
arXiv Detail & Related papers (2021-06-14T05:39:09Z) - We Need to Talk About Random Splits [3.236124102160291]
Gorman and Bedrick argued for using random splits rather than standard splits in NLP experiments.
We argue that random splits, like standard splits, lead to overly optimistic performance estimates.
arXiv Detail & Related papers (2020-05-01T22:14:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.