Predictive Modeling in the Presence of Nuisance-Induced Spurious
Correlations
- URL: http://arxiv.org/abs/2107.00520v1
- Date: Tue, 29 Jun 2021 18:12:59 GMT
- Title: Predictive Modeling in the Presence of Nuisance-Induced Spurious
Correlations
- Authors: Aahlad Puli, Lily H. Zhang, Eric K. Oermann, Rajesh Ranganath
- Abstract summary: In classification tasks, spurious correlations are induced by a changing relationship between the label and some nuisance variables.
We formalize a family of distributions that only differ in the nuisance-label relationship.
We produce models that predict pneumonia under strong spurious correlations.
- Score: 18.529899583515206
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep predictive models often make use of spurious correlations between the
label and the covariates that differ between training and test distributions.
In many classification tasks, spurious correlations are induced by a changing
relationship between the label and some nuisance variables correlated with the
covariates. For example, in classifying animals in natural images, the
background, which is the nuisance, can predict the type of animal, but this
nuisance label relationship does not always hold. This nuisance-label
relationship does not always hold. We formalize a family of distributions that
only differ in the nuisance-label relationship and and introduce a distribution
where this relationship is broken called the nuisance-randomized distribution.
We introduce a set of predictive models built from the nuisance-randomized
distribution with representations, that when conditioned on, do not correlate
the label and the nuisance. For models in this set, we lower bound the
performance for any member of the family with the mutual information between
the representation and the label under the nuisance-randomized distribution. To
build predictive models that maximize the performance lower bound, we develop
Nuisance-Randomized Distillation (NURD). We evaluate NURD on a synthetic
example, colored-MNIST, and classifying chest X-rays. When using non-lung
patches as the nuisance in classifying chest X-rays, NURD produces models that
predict pneumonia under strong spurious correlations.
Related papers
- Mitigating Spurious Correlations via Disagreement Probability [4.8884049398279705]
Models trained with empirical risk minimization (ERM) are prone to be biased towards spurious correlations between target labels and bias attributes.
We introduce a training objective designed to robustly enhance model performance across all data samples.
We then derive a debiasing method, Disagreement Probability based Resampling for debiasing (DPR), which does not require bias labels.
arXiv Detail & Related papers (2024-11-04T02:44:04Z) - Towards Robust Text Classification: Mitigating Spurious Correlations with Causal Learning [2.7813683000222653]
We propose the Causally Calibrated Robust ( CCR) to reduce models' reliance on spurious correlations.
CCR integrates a causal feature selection method based on counterfactual reasoning, along with an inverse propensity weighting (IPW) loss function.
We show that CCR state-of-the-art performance among methods without group labels, and in some cases, it can compete with the models that utilize group labels.
arXiv Detail & Related papers (2024-11-01T21:29:07Z) - Theory on Score-Mismatched Diffusion Models and Zero-Shot Conditional Samplers [49.97755400231656]
We present the first performance guarantee with explicit dimensional general score-mismatched diffusion samplers.
We show that score mismatches result in an distributional bias between the target and sampling distributions, proportional to the accumulated mismatch between the target and training distributions.
This result can be directly applied to zero-shot conditional samplers for any conditional model, irrespective of measurement noise.
arXiv Detail & Related papers (2024-10-17T16:42:12Z) - A Note on the Convergence of Denoising Diffusion Probabilistic Models [3.1767625261233046]
We derive a quantitative upper bound on the Wasserstein distance between the data-generating distribution and the distribution learned by a diffusion model.
Unlike previous works in this field, our result does not make assumptions on the learned score function.
arXiv Detail & Related papers (2023-12-10T20:29:58Z) - Stubborn Lexical Bias in Data and Models [50.79738900885665]
We use a new statistical method to examine whether spurious patterns in data appear in models trained on the data.
We apply an optimization approach to *reweight* the training data, reducing thousands of spurious correlations.
Surprisingly, though this method can successfully reduce lexical biases in the training data, we still find strong evidence of corresponding bias in the trained models.
arXiv Detail & Related papers (2023-06-03T20:12:27Z) - Nonparametric Identifiability of Causal Representations from Unknown
Interventions [63.1354734978244]
We study causal representation learning, the task of inferring latent causal variables and their causal relations from mixtures of the variables.
Our goal is to identify both the ground truth latents and their causal graph up to a set of ambiguities which we show to be irresolvable from interventional data.
arXiv Detail & Related papers (2023-06-01T10:51:58Z) - Breaking the Spurious Causality of Conditional Generation via Fairness
Intervention with Corrective Sampling [77.15766509677348]
Conditional generative models often inherit spurious correlations from the training dataset.
This can result in label-conditional distributions that are imbalanced with respect to another latent attribute.
We propose a general two-step strategy to mitigate this issue.
arXiv Detail & Related papers (2022-12-05T08:09:33Z) - Nuisances via Negativa: Adjusting for Spurious Correlations via Data Augmentation [32.66196135141696]
Features with varying relationships to the label are nuisances.
Models that exploit nuisance-label relationships face performance degradation when these relationships change.
We develop an approach to use knowledge about the semantics by corrupting them in data.
arXiv Detail & Related papers (2022-10-04T01:40:31Z) - Adversarial Robustness through the Lens of Causality [105.51753064807014]
adversarial vulnerability of deep neural networks has attracted significant attention in machine learning.
We propose to incorporate causality into mitigating adversarial vulnerability.
Our method can be seen as the first attempt to leverage causality for mitigating adversarial vulnerability.
arXiv Detail & Related papers (2021-06-11T06:55:02Z) - Towards Robust Classification Model by Counterfactual and Invariant Data
Generation [7.488317734152585]
Spuriousness occurs when some features correlate with labels but are not causal.
We propose two data generation processes to reduce spuriousness.
Our data generations outperform state-of-the-art methods in accuracy when spurious correlations break.
arXiv Detail & Related papers (2021-06-02T12:48:29Z) - When Hearst Is not Enough: Improving Hypernymy Detection from Corpus
with Distributional Models [59.46552488974247]
This paper addresses whether an is-a relationship exists between words (x, y) with the help of large textual corpora.
Recent studies suggest that pattern-based ones are superior, if large-scale Hearst pairs are extracted and fed, with the sparsity of unseen (x, y) pairs relieved.
For the first time, this paper quantifies the non-negligible existence of those specific cases. We also demonstrate that distributional methods are ideal to make up for pattern-based ones in such cases.
arXiv Detail & Related papers (2020-10-10T08:34:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.