Learning under random distributional shifts
- URL: http://arxiv.org/abs/2306.02948v2
- Date: Mon, 30 Oct 2023 05:27:51 GMT
- Title: Learning under random distributional shifts
- Authors: Kirk Bansak, Elisabeth Paulson, Dominik Rothenh\"ausler
- Abstract summary: We consider a class of random distribution shift models that capture arbitrary changes in the underlying covariate space.
We show that the hybrid approach is robust to the strength of the distribution shift and the proxy relationship.
In two high-impact domains, we find that the proposed approach results in substantially lower mean-squared error.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Many existing approaches for generating predictions in settings with
distribution shift model distribution shifts as adversarial or low-rank in
suitable representations. In various real-world settings, however, we might
expect shifts to arise through the superposition of many small and random
changes in the population and environment. Thus, we consider a class of random
distribution shift models that capture arbitrary changes in the underlying
covariate space, and dense, random shocks to the relationship between the
covariates and the outcomes. In this setting, we characterize the benefits and
drawbacks of several alternative prediction strategies: the standard approach
that directly predicts the long-term outcome of interest, the proxy approach
that directly predicts a shorter-term proxy outcome, and a hybrid approach that
utilizes both the long-term policy outcome and (shorter-term) proxy outcome(s).
We show that the hybrid approach is robust to the strength of the distribution
shift and the proxy relationship. We apply this method to datasets in two
high-impact domains: asylum-seeker assignment and early childhood education. In
both settings, we find that the proposed approach results in substantially
lower mean-squared error than current approaches.
Related papers
- Probabilistic Conformal Prediction with Approximate Conditional Validity [81.30551968980143]
We develop a new method for generating prediction sets that combines the flexibility of conformal methods with an estimate of the conditional distribution.
Our method consistently outperforms existing approaches in terms of conditional coverage.
arXiv Detail & Related papers (2024-07-01T20:44:48Z) - TransFusion: Covariate-Shift Robust Transfer Learning for High-Dimensional Regression [11.040033344386366]
We propose a two-step method with a novel fused-regularizer to improve the learning performance on a target task with limited samples.
Nonasymptotic bound is provided for the estimation error of the target model.
We extend the method to a distributed setting, allowing for a pretraining-finetuning strategy.
arXiv Detail & Related papers (2024-04-01T14:58:16Z) - Proxy Methods for Domain Adaptation [78.03254010884783]
proxy variables allow for adaptation to distribution shift without explicitly recovering or modeling latent variables.
We develop a two-stage kernel estimation approach to adapt to complex distribution shifts in both settings.
arXiv Detail & Related papers (2024-03-12T09:32:41Z) - Uncertainty Quantification via Stable Distribution Propagation [60.065272548502]
We propose a new approach for propagating stable probability distributions through neural networks.
Our method is based on local linearization, which we show to be an optimal approximation in terms of total variation distance for the ReLU non-linearity.
arXiv Detail & Related papers (2024-02-13T09:40:19Z) - Causality-oriented robustness: exploiting general additive interventions [3.871660145364189]
In this paper, we focus on causality-oriented robustness and propose Distributional Robustness via Invariant Gradients (DRIG)
In a linear setting, we prove that DRIG yields predictions that are robust among a data-dependent class of distribution shifts.
We extend our approach to the semi-supervised domain adaptation setting to further improve prediction performance.
arXiv Detail & Related papers (2023-07-18T16:22:50Z) - Supervised Contrastive Learning with Heterogeneous Similarity for
Distribution Shifts [3.7819322027528113]
We propose a new regularization using the supervised contrastive learning to prevent such overfitting and to train models that do not degrade their performance under the distribution shifts.
Experiments on benchmark datasets that emulate distribution shifts, including subpopulation shift and domain generalization, demonstrate the advantage of the proposed method.
arXiv Detail & Related papers (2023-04-07T01:45:09Z) - Sinkhorn-Flow: Predicting Probability Mass Flow in Dynamical Systems
Using Optimal Transport [89.61692654941106]
We propose a new approach to predicting such mass flow over time using optimal transport.
We apply our approach to the task of predicting how communities will evolve over time in social network settings.
arXiv Detail & Related papers (2023-03-14T07:25:44Z) - Variance-Reduced Heterogeneous Federated Learning via Stratified Client
Selection [31.401919362978017]
We propose a novel stratified client selection scheme to reduce the variance for the pursuit of better convergence and higher accuracy.
We present an optimized sample size allocation scheme by considering the diversity of stratum's variability.
Experimental results confirm that our approach not only allows for better performance relative to state-of-the-art methods but also is compatible with prevalent FL algorithms.
arXiv Detail & Related papers (2022-01-15T05:41:36Z) - CovarianceNet: Conditional Generative Model for Correct Covariance
Prediction in Human Motion Prediction [71.31516599226606]
We present a new method to correctly predict the uncertainty associated with the predicted distribution of future trajectories.
Our approach, CovariaceNet, is based on a Conditional Generative Model with Gaussian latent variables.
arXiv Detail & Related papers (2021-09-07T09:38:24Z) - Robustness via Cross-Domain Ensembles [0.5801044612920816]
We present a method for making neural network predictions robust to shifts from the training data distribution.
The proposed method is based on making predictions via a diverse set of cues and ensembling them into one strong prediction.
arXiv Detail & Related papers (2021-03-19T17:28:03Z) - A One-step Approach to Covariate Shift Adaptation [82.01909503235385]
A default assumption in many machine learning scenarios is that the training and test samples are drawn from the same probability distribution.
We propose a novel one-step approach that jointly learns the predictive model and the associated weights in one optimization.
arXiv Detail & Related papers (2020-07-08T11:35:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.