Uncertainty in Contrastive Learning: On the Predictability of Downstream
Performance
- URL: http://arxiv.org/abs/2207.09336v1
- Date: Tue, 19 Jul 2022 15:44:59 GMT
- Title: Uncertainty in Contrastive Learning: On the Predictability of Downstream
Performance
- Authors: Shervin Ardeshir and Navid Azizan
- Abstract summary: We study whether the uncertainty of such a representation can be quantified for a single datapoint in a meaningful way.
We show that this goal can be achieved by directly estimating the distribution of the training data in the embedding space.
- Score: 7.411571833582691
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The superior performance of some of today's state-of-the-art deep learning
models is to some extent owed to extensive (self-)supervised contrastive
pretraining on large-scale datasets. In contrastive learning, the network is
presented with pairs of positive (similar) and negative (dissimilar) datapoints
and is trained to find an embedding vector for each datapoint, i.e., a
representation, which can be further fine-tuned for various downstream tasks.
In order to safely deploy these models in critical decision-making systems, it
is crucial to equip them with a measure of their uncertainty or reliability.
However, due to the pairwise nature of training a contrastive model, and the
lack of absolute labels on the output (an abstract embedding vector), adapting
conventional uncertainty estimation techniques to such models is non-trivial.
In this work, we study whether the uncertainty of such a representation can be
quantified for a single datapoint in a meaningful way. In other words, we
explore if the downstream performance on a given datapoint is predictable,
directly from its pre-trained embedding. We show that this goal can be achieved
by directly estimating the distribution of the training data in the embedding
space and accounting for the local consistency of the representations. Our
experiments show that this notion of uncertainty for an embedding vector often
strongly correlates with its downstream accuracy.
Related papers
- Downstream-Pretext Domain Knowledge Traceback for Active Learning [138.02530777915362]
We propose a downstream-pretext domain knowledge traceback (DOKT) method that traces the data interactions of downstream knowledge and pre-training guidance.
DOKT consists of a traceback diversity indicator and a domain-based uncertainty estimator.
Experiments conducted on ten datasets show that our model outperforms other state-of-the-art methods.
arXiv Detail & Related papers (2024-07-20T01:34:13Z) - Quantifying Representation Reliability in Self-Supervised Learning Models [12.485580780944083]
Self-supervised learning models extract general-purpose representations from data.
We introduce a formal definition of representation reliability.
We propose an ensemble-based method for estimating the representation reliability without knowing the downstream tasks a priori.
arXiv Detail & Related papers (2023-05-31T21:57:33Z) - Certifying Data-Bias Robustness in Linear Regression [12.00314910031517]
We present a technique for certifying whether linear regression models are pointwise-robust to label bias in a training dataset.
We show how to solve this problem exactly for individual test points, and provide an approximate but more scalable method.
We also unearth gaps in bias-robustness, such as high levels of non-robustness for certain bias assumptions on some datasets.
arXiv Detail & Related papers (2022-06-07T20:47:07Z) - Robust Flow-based Conformal Inference (FCI) with Statistical Guarantee [4.821312633849745]
We develop a series of conformal inference methods, including building predictive sets and inferring outliers for complex and high-dimensional data.
We evaluate our method, robust flow-based conformal inference, on benchmark datasets.
arXiv Detail & Related papers (2022-05-22T04:17:30Z) - Leveraging Unlabeled Data to Predict Out-of-Distribution Performance [63.740181251997306]
Real-world machine learning deployments are characterized by mismatches between the source (training) and target (test) distributions.
In this work, we investigate methods for predicting the target domain accuracy using only labeled source data and unlabeled target data.
We propose Average Thresholded Confidence (ATC), a practical method that learns a threshold on the model's confidence, predicting accuracy as the fraction of unlabeled examples.
arXiv Detail & Related papers (2022-01-11T23:01:12Z) - Evaluating Predictive Uncertainty and Robustness to Distributional Shift
Using Real World Data [0.0]
We propose metrics for general regression tasks using the Shifts Weather Prediction dataset.
We also present an evaluation of the baseline methods using these metrics.
arXiv Detail & Related papers (2021-11-08T17:32:10Z) - Accurate and Robust Feature Importance Estimation under Distribution
Shifts [49.58991359544005]
PRoFILE is a novel feature importance estimation method.
We show significant improvements over state-of-the-art approaches, both in terms of fidelity and robustness.
arXiv Detail & Related papers (2020-09-30T05:29:01Z) - Graph Embedding with Data Uncertainty [113.39838145450007]
spectral-based subspace learning is a common data preprocessing step in many machine learning pipelines.
Most subspace learning methods do not take into consideration possible measurement inaccuracies or artifacts that can lead to data with high uncertainty.
arXiv Detail & Related papers (2020-09-01T15:08:23Z) - Learning while Respecting Privacy and Robustness to Distributional
Uncertainties and Adversarial Data [66.78671826743884]
The distributionally robust optimization framework is considered for training a parametric model.
The objective is to endow the trained model with robustness against adversarially manipulated input data.
Proposed algorithms offer robustness with little overhead.
arXiv Detail & Related papers (2020-07-07T18:25:25Z) - Unlabelled Data Improves Bayesian Uncertainty Calibration under
Covariate Shift [100.52588638477862]
We develop an approximate Bayesian inference scheme based on posterior regularisation.
We demonstrate the utility of our method in the context of transferring prognostic models of prostate cancer across globally diverse populations.
arXiv Detail & Related papers (2020-06-26T13:50:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.