Enhancing Robustness of Foundation Model Representations under
Provenance-related Distribution Shifts
- URL: http://arxiv.org/abs/2312.05435v1
- Date: Sat, 9 Dec 2023 02:02:45 GMT
- Title: Enhancing Robustness of Foundation Model Representations under
Provenance-related Distribution Shifts
- Authors: Xiruo Ding, Zhecheng Sheng, Brian Hur, Feng Chen, Serguei V. S.
Pakhomov, Trevor Cohen
- Abstract summary: We examine the stability of models based on foundation models under distribution shift.
We focus on confounding by provenance, a form of distribution shift that emerges in the context of multi-institutional datasets.
Results indicate that while foundation models do show some out-of-the-box robustness to confounding-by-provenance related distribution shifts, this can be improved through adjustment.
- Score: 8.298173603769063
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Foundation models are a current focus of attention in both industry and
academia. While they have shown their capabilities in a variety of tasks,
in-depth research is required to determine their robustness to distribution
shift when used as a basis for supervised machine learning. This is especially
important in the context of clinical data, with particular limitations related
to data accessibility, lack of pretraining materials, and limited availability
of high-quality annotations. In this work, we examine the stability of models
based on representations from foundation models under distribution shift. We
focus on confounding by provenance, a form of distribution shift that emerges
in the context of multi-institutional datasets when there are differences in
source-specific language use and class distributions. Using a sampling strategy
that synthetically induces varying degrees of distribution shift, we evaluate
the extent to which representations from foundation models result in
predictions that are inherently robust to confounding by provenance.
Additionally, we examine the effectiveness of a straightforward confounding
adjustment method inspired by Pearl's conception of backdoor adjustment.
Results indicate that while foundation models do show some out-of-the-box
robustness to confounding-by-provenance related distribution shifts, this can
be considerably improved through adjustment. These findings suggest a need for
deliberate adjustment of predictive models using representations from
foundation models in the context of source-specific distributional differences.
Related papers
- Quantifying Distribution Shifts and Uncertainties for Enhanced Model Robustness in Machine Learning Applications [0.0]
This study explores model adaptation and generalization by utilizing synthetic data.
We employ quantitative measures such as Kullback-Leibler divergence, Jensen-Shannon distance, and Mahalanobis distance to assess data similarity.
Our findings suggest that utilizing statistical measures, such as the Mahalanobis distance, to determine whether model predictions fall within the low-error "interpolation regime" or the high-error "extrapolation regime" provides a complementary method for assessing distribution shift and model uncertainty.
arXiv Detail & Related papers (2024-05-03T10:05:31Z) - Identifiable Latent Neural Causal Models [82.14087963690561]
Causal representation learning seeks to uncover latent, high-level causal representations from low-level observed data.
We determine the types of distribution shifts that do contribute to the identifiability of causal representations.
We translate our findings into a practical algorithm, allowing for the acquisition of reliable latent causal representations.
arXiv Detail & Related papers (2024-03-23T04:13:55Z) - Cross-modality debiasing: using language to mitigate sub-population shifts in imaging [28.88097536026781]
Sub-population shift accounts for a significant source of algorithmic bias and calls for distributional robustness.
Recent studies found inherent distributional robustness in multi-modality foundation models, such as the vision-language model CLIP.
We propose leveraging natural language inputs to debias the image feature representations, to improve worst-case performance on sub-populations.
arXiv Detail & Related papers (2024-02-02T18:54:48Z) - On the Out of Distribution Robustness of Foundation Models in Medical
Image Segmentation [47.95611203419802]
Foundations for vision and language, pre-trained on extensive sets of natural image and text data, have emerged as a promising approach.
We compare the generalization performance to unseen domains of various pre-trained models after being fine-tuned on the same in-distribution dataset.
We further developed a new Bayesian uncertainty estimation for frozen models and used them as an indicator to characterize the model's performance on out-of-distribution data.
arXiv Detail & Related papers (2023-11-18T14:52:10Z) - Context matters for fairness -- a case study on the effect of spatial
distribution shifts [10.351739012146378]
We present a case study on the newly released American Census datasets.
We show how remarkably can spatial distribution shifts affect predictive- and fairness-related performance of a model.
Our study suggests that robustness to distribution shifts is necessary before deploying a model to another context.
arXiv Detail & Related papers (2022-06-23T01:09:46Z) - Certifying Model Accuracy under Distribution Shifts [151.67113334248464]
We present provable robustness guarantees on the accuracy of a model under bounded Wasserstein shifts of the data distribution.
We show that a simple procedure that randomizes the input of the model within a transformation space is provably robust to distributional shifts under the transformation.
arXiv Detail & Related papers (2022-01-28T22:03:50Z) - Accuracy on the Line: On the Strong Correlation Between
Out-of-Distribution and In-Distribution Generalization [89.73665256847858]
We show that out-of-distribution performance is strongly correlated with in-distribution performance for a wide range of models and distribution shifts.
Specifically, we demonstrate strong correlations between in-distribution and out-of-distribution performance on variants of CIFAR-10 & ImageNet.
We also investigate cases where the correlation is weaker, for instance some synthetic distribution shifts from CIFAR-10-C and the tissue classification dataset Camelyon17-WILDS.
arXiv Detail & Related papers (2021-07-09T19:48:23Z) - Predicting with Confidence on Unseen Distributions [90.68414180153897]
We connect domain adaptation and predictive uncertainty literature to predict model accuracy on challenging unseen distributions.
We find that the difference of confidences (DoC) of a classifier's predictions successfully estimates the classifier's performance change over a variety of shifts.
We specifically investigate the distinction between synthetic and natural distribution shifts and observe that despite its simplicity DoC consistently outperforms other quantifications of distributional difference.
arXiv Detail & Related papers (2021-07-07T15:50:18Z) - Accurate and Robust Feature Importance Estimation under Distribution
Shifts [49.58991359544005]
PRoFILE is a novel feature importance estimation method.
We show significant improvements over state-of-the-art approaches, both in terms of fidelity and robustness.
arXiv Detail & Related papers (2020-09-30T05:29:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.