Domain Adaptation under Missingness Shift
- URL: http://arxiv.org/abs/2211.02093v3
- Date: Wed, 3 May 2023 20:38:36 GMT
- Title: Domain Adaptation under Missingness Shift
- Authors: Helen Zhou, Sivaraman Balakrishnan, Zachary C. Lipton
- Abstract summary: We introduce the problem of Domain Adaptation under Missingness Shift (DAMS)
Rates of missing data often depend on record-keeping policies and thus may change across times and locations.
In experiments on synthetic and semi-synthetic data, we demonstrate the promise of our methods when assumptions hold.
- Score: 38.650099178537864
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Rates of missing data often depend on record-keeping policies and thus may
change across times and locations, even when the underlying features are
comparatively stable. In this paper, we introduce the problem of Domain
Adaptation under Missingness Shift (DAMS). Here, (labeled) source data and
(unlabeled) target data would be exchangeable but for different missing data
mechanisms. We show that if missing data indicators are available, DAMS reduces
to covariate shift. Addressing cases where such indicators are absent, we
establish the following theoretical results for underreporting completely at
random: (i) covariate shift is violated (adaptation is required); (ii) the
optimal linear source predictor can perform arbitrarily worse on the target
domain than always predicting the mean; (iii) the optimal target predictor can
be identified, even when the missingness rates themselves are not; and (iv) for
linear models, a simple analytic adjustment yields consistent estimates of the
optimal target parameters. In experiments on synthetic and semi-synthetic data,
we demonstrate the promise of our methods when assumptions hold. Finally, we
discuss a rich family of future extensions.
Related papers
- Adapting to Shifting Correlations with Unlabeled Data Calibration [6.84735357291896]
Distribution shifts between sites can seriously degrade model performance since models are prone to exploiting unstable correlations.
We propose Generalized Prevalence Adjustment (GPA), a flexible method that adjusts model predictions to the shifting correlations between prediction target and confounders.
GPA can infer the interaction between target and confounders in new sites using unlabeled samples from those sites.
arXiv Detail & Related papers (2024-09-09T18:45:43Z) - Learning When the Concept Shifts: Confounding, Invariance, and Dimension Reduction [5.38274042816001]
In observational data, the distribution shift is often driven by unobserved confounding factors.
This motivates us to study the domain adaptation problem with observational data.
We show a model that uses the learned lower-dimensional subspace can incur nearly ideal gap between target and source risk.
arXiv Detail & Related papers (2024-06-22T17:43:08Z) - Conformal Prediction with Missing Values [19.18178194789968]
We first show that the marginal coverage guarantee of conformal prediction holds on imputed data for any missingness distribution.
We then show that a universally consistent quantile regression algorithm trained on the imputed data is Bayes optimal for the pinball risk.
arXiv Detail & Related papers (2023-06-05T09:28:03Z) - Adapting to Latent Subgroup Shifts via Concepts and Proxies [82.01141290360562]
We show that the optimal target predictor can be non-parametrically identified with the help of concept and proxy variables available only in the source domain.
For continuous observations, we propose a latent variable model specific to the data generation process at hand.
arXiv Detail & Related papers (2022-12-21T18:30:22Z) - Uncertainty-guided Source-free Domain Adaptation [77.3844160723014]
Source-free domain adaptation (SFDA) aims to adapt a classifier to an unlabelled target data set by only using a pre-trained source model.
We propose quantifying the uncertainty in the source model predictions and utilizing it to guide the target adaptation.
arXiv Detail & Related papers (2022-08-16T08:03:30Z) - Leveraging Unlabeled Data to Predict Out-of-Distribution Performance [63.740181251997306]
Real-world machine learning deployments are characterized by mismatches between the source (training) and target (test) distributions.
In this work, we investigate methods for predicting the target domain accuracy using only labeled source data and unlabeled target data.
We propose Average Thresholded Confidence (ATC), a practical method that learns a threshold on the model's confidence, predicting accuracy as the fraction of unlabeled examples.
arXiv Detail & Related papers (2022-01-11T23:01:12Z) - BCD Nets: Scalable Variational Approaches for Bayesian Causal Discovery [97.79015388276483]
A structural equation model (SEM) is an effective framework to reason over causal relationships represented via a directed acyclic graph (DAG)
Recent advances enabled effective maximum-likelihood point estimation of DAGs from observational data.
We propose BCD Nets, a variational framework for estimating a distribution over DAGs characterizing a linear-Gaussian SEM.
arXiv Detail & Related papers (2021-12-06T03:35:21Z) - Unlabelled Data Improves Bayesian Uncertainty Calibration under
Covariate Shift [100.52588638477862]
We develop an approximate Bayesian inference scheme based on posterior regularisation.
We demonstrate the utility of our method in the context of transferring prognostic models of prostate cancer across globally diverse populations.
arXiv Detail & Related papers (2020-06-26T13:50:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.