Learning Invariant Representations with Missing Data
- URL: http://arxiv.org/abs/2112.00881v1
- Date: Wed, 1 Dec 2021 23:14:34 GMT
- Title: Learning Invariant Representations with Missing Data
- Authors: Mark Goldstein, J\"orn-Henrik Jacobsen, Olina Chau, Adriel Saporta,
Aahlad Puli, Rajesh Ranganath, Andrew C. Miller
- Abstract summary: Models that satisfy particular independencies involving correlation-inducing textitnuisance variables have guarantees on their test performance.
We derive acrshortmmd estimators used for invariance objectives under missing nuisances.
On simulations and clinical data, optimizing through these estimates achieves test performance similar to using estimators that make use of the full data.
- Score: 18.307438471163774
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Spurious correlations allow flexible models to predict well during training
but poorly on related test populations. Recent work has shown that models that
satisfy particular independencies involving correlation-inducing
\textit{nuisance} variables have guarantees on their test performance.
Enforcing such independencies requires nuisances to be observed during
training. However, nuisances, such as demographics or image background labels,
are often missing. Enforcing independence on just the observed data does not
imply independence on the entire population. Here we derive \acrshort{mmd}
estimators used for invariance objectives under missing nuisances. On
simulations and clinical data, optimizing through these estimates achieves test
performance similar to using estimators that make use of the full data.
Related papers
- On Different Notions of Redundancy in Conditional-Independence-Based Discovery of Graphical Models [11.384020227038961]
We show that due to the conciseness of the graphical representation, there are often many tests that are not used in the construction of the graph.
We show that not all tests contain this additional information and that such redundant tests have to be applied with care.
arXiv Detail & Related papers (2025-02-12T16:08:48Z) - Measuring and Improving Attentiveness to Partial Inputs with Counterfactuals [91.59906995214209]
We propose a new evaluation method, Counterfactual Attentiveness Test (CAT)
CAT uses counterfactuals by replacing part of the input with its counterpart from a different example, expecting an attentive model to change its prediction.
We show that GPT3 becomes less attentive with an increased number of demonstrations, while its accuracy on the test data improves.
arXiv Detail & Related papers (2023-11-16T06:27:35Z) - Stubborn Lexical Bias in Data and Models [50.79738900885665]
We use a new statistical method to examine whether spurious patterns in data appear in models trained on the data.
We apply an optimization approach to *reweight* the training data, reducing thousands of spurious correlations.
Surprisingly, though this method can successfully reduce lexical biases in the training data, we still find strong evidence of corresponding bias in the trained models.
arXiv Detail & Related papers (2023-06-03T20:12:27Z) - Improving Adaptive Conformal Prediction Using Self-Supervised Learning [72.2614468437919]
We train an auxiliary model with a self-supervised pretext task on top of an existing predictive model and use the self-supervised error as an additional feature to estimate nonconformity scores.
We empirically demonstrate the benefit of the additional information using both synthetic and real data on the efficiency (width), deficit, and excess of conformal prediction intervals.
arXiv Detail & Related papers (2023-02-23T18:57:14Z) - Sharing pattern submodels for prediction with missing values [12.981974894538668]
Missing values are unavoidable in many applications of machine learning and present challenges both during training and at test time.
We propose an alternative approach, called sharing pattern submodels, which i) makes predictions robust to missing values at test time, ii) maintains or improves the predictive power of pattern submodels andiii) has a short description, enabling improved interpretability.
arXiv Detail & Related papers (2022-06-22T15:09:40Z) - Statistical Estimation from Dependent Data [37.73584699735133]
We consider a general statistical estimation problem wherein binary labels across different observations are not independent conditioned on their feature vectors.
We model these dependencies in the language of Markov Random Fields.
We provide algorithms and statistically efficient estimation rates for this model.
arXiv Detail & Related papers (2021-07-20T21:18:06Z) - On Disentangled Representations Learned From Correlated Data [59.41587388303554]
We bridge the gap to real-world scenarios by analyzing the behavior of the most prominent disentanglement approaches on correlated data.
We show that systematically induced correlations in the dataset are being learned and reflected in the latent representations.
We also demonstrate how to resolve these latent correlations, either using weak supervision during training or by post-hoc correcting a pre-trained model with a small number of labels.
arXiv Detail & Related papers (2020-06-14T12:47:34Z) - Performance metrics for intervention-triggering prediction models do not
reflect an expected reduction in outcomes from using the model [71.9860741092209]
Clinical researchers often select among and evaluate risk prediction models.
Standard metrics calculated from retrospective data are only related to model utility under certain assumptions.
When predictions are delivered repeatedly throughout time, the relationship between standard metrics and utility is further complicated.
arXiv Detail & Related papers (2020-06-02T16:26:49Z) - TraDE: Transformers for Density Estimation [101.20137732920718]
TraDE is a self-attention-based architecture for auto-regressive density estimation.
We present a suite of tasks such as regression using generated samples, out-of-distribution detection, and robustness to noise in the training data.
arXiv Detail & Related papers (2020-04-06T07:32:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.