Nuisances via Negativa: Adjusting for Spurious Correlations via Data Augmentation
- URL: http://arxiv.org/abs/2210.01302v3
- Date: Wed, 3 Jul 2024 08:06:56 GMT
- Title: Nuisances via Negativa: Adjusting for Spurious Correlations via Data Augmentation
- Authors: Aahlad Puli, Nitish Joshi, Yoav Wald, He He, Rajesh Ranganath,
- Abstract summary: Features with varying relationships to the label are nuisances.
Models that exploit nuisance-label relationships face performance degradation when these relationships change.
We develop an approach to use knowledge about the semantics by corrupting them in data.
- Score: 32.66196135141696
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In prediction tasks, there exist features that are related to the label in the same way across different settings for that task; these are semantic features or semantics. Features with varying relationships to the label are nuisances. For example, in detecting cows from natural images, the shape of the head is semantic but because images of cows often have grass backgrounds but not always, the background is a nuisance. Models that exploit nuisance-label relationships face performance degradation when these relationships change. Building models robust to such changes requires additional knowledge beyond samples of the features and labels. For example, existing work uses annotations of nuisances or assumes ERM-trained models depend on nuisances. Approaches to integrate new kinds of additional knowledge enlarge the settings where robust models can be built. We develop an approach to use knowledge about the semantics by corrupting them in data, and then using the corrupted data to produce models which identify correlations between nuisances and the label. Once these correlations are identified, they can be used to adjust for where nuisances drive predictions. We study semantic corruptions in powering different spurious-correlation avoiding methods on multiple out-of-distribution (OOD) tasks like classifying waterbirds, natural language inference (NLI), and detecting cardiomegaly in chest X-rays.
Related papers
- Common-Sense Bias Discovery and Mitigation for Classification Tasks [16.8259488742528]
We propose a framework to extract feature clusters in a dataset based on image descriptions.
The analyzed features and correlations are human-interpretable, so we name the method Common-Sense Bias Discovery (CSBD)
Experiments show that our method discovers novel biases on multiple classification tasks for two benchmark image datasets.
arXiv Detail & Related papers (2024-01-24T03:56:07Z) - Improving Fairness using Vision-Language Driven Image Augmentation [60.428157003498995]
Fairness is crucial when training a deep-learning discriminative model, especially in the facial domain.
Models tend to correlate specific characteristics (such as age and skin color) with unrelated attributes (downstream tasks)
This paper proposes a method to mitigate these correlations to improve fairness.
arXiv Detail & Related papers (2023-11-02T19:51:10Z) - Stubborn Lexical Bias in Data and Models [50.79738900885665]
We use a new statistical method to examine whether spurious patterns in data appear in models trained on the data.
We apply an optimization approach to *reweight* the training data, reducing thousands of spurious correlations.
Surprisingly, though this method can successfully reduce lexical biases in the training data, we still find strong evidence of corresponding bias in the trained models.
arXiv Detail & Related papers (2023-06-03T20:12:27Z) - Label-Retrieval-Augmented Diffusion Models for Learning from Noisy
Labels [61.97359362447732]
Learning from noisy labels is an important and long-standing problem in machine learning for real applications.
In this paper, we reformulate the label-noise problem from a generative-model perspective.
Our model achieves new state-of-the-art (SOTA) results on all the standard real-world benchmark datasets.
arXiv Detail & Related papers (2023-05-31T03:01:36Z) - Causal Transportability for Visual Recognition [70.13627281087325]
We show that standard classifiers fail because the association between images and labels is not transportable across settings.
We then show that the causal effect, which severs all sources of confounding, remains invariant across domains.
This motivates us to develop an algorithm to estimate the causal effect for image classification.
arXiv Detail & Related papers (2022-04-26T15:02:11Z) - Nuisance-Label Supervision: Robustness Improvement by Free Labels [14.711384503643995]
We present a Nuisance-label Supervision (NLS) module, which can make models more robust to nuisance factor variations.
Experiments show consistent improvement in robustness towards image corruption and appearance change in action recognition.
arXiv Detail & Related papers (2021-10-14T02:07:00Z) - Predictive Modeling in the Presence of Nuisance-Induced Spurious
Correlations [18.529899583515206]
In classification tasks, spurious correlations are induced by a changing relationship between the label and some nuisance variables.
We formalize a family of distributions that only differ in the nuisance-label relationship.
We produce models that predict pneumonia under strong spurious correlations.
arXiv Detail & Related papers (2021-06-29T18:12:59Z) - Towards Robust Classification Model by Counterfactual and Invariant Data
Generation [7.488317734152585]
Spuriousness occurs when some features correlate with labels but are not causal.
We propose two data generation processes to reduce spuriousness.
Our data generations outperform state-of-the-art methods in accuracy when spurious correlations break.
arXiv Detail & Related papers (2021-06-02T12:48:29Z) - Learning to Model and Ignore Dataset Bias with Mixed Capacity Ensembles [66.15398165275926]
We propose a method that can automatically detect and ignore dataset-specific patterns, which we call dataset biases.
Our method trains a lower capacity model in an ensemble with a higher capacity model.
We show improvement in all settings, including a 10 point gain on the visual question answering dataset.
arXiv Detail & Related papers (2020-11-07T22:20:03Z) - Learning Causal Models Online [103.87959747047158]
Predictive models can rely on spurious correlations in the data for making predictions.
One solution for achieving strong generalization is to incorporate causal structures in the models.
We propose an online algorithm that continually detects and removes spurious features.
arXiv Detail & Related papers (2020-06-12T20:49:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.