Enhancing Distributional Stability among Sub-populations
- URL: http://arxiv.org/abs/2206.02990v2
- Date: Wed, 14 Feb 2024 02:26:15 GMT
- Title: Enhancing Distributional Stability among Sub-populations
- Authors: Jiashuo Liu, Jiayun Wu, Jie Peng, Xiaoyu Wu, Yang Zheng, Bo Li, Peng
Cui
- Abstract summary: Enhancing the stability of machine learning algorithms under distributional shifts is at the heart of the Out-of-Distribution (OOD) Generalization problem.
We propose a novel stable risk minimization (SRM) algorithm to enhance the model's stability w.r.t.
Experimental results are consistent with our intuition and validate the effectiveness of our algorithm.
- Score: 32.66329730287957
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Enhancing the stability of machine learning algorithms under distributional
shifts is at the heart of the Out-of-Distribution (OOD) Generalization problem.
Derived from causal learning, recent works of invariant learning pursue strict
invariance with multiple training environments. Although intuitively
reasonable, strong assumptions on the availability and quality of environments
are made to learn the strict invariance property. In this work, we come up with
the ``distributional stability" notion to mitigate such limitations. It
quantifies the stability of prediction mechanisms among sub-populations down to
a prescribed scale. Based on this, we propose the learnability assumption and
derive the generalization error bound under distribution shifts. Inspired by
theoretical analyses, we propose our novel stable risk minimization (SRM)
algorithm to enhance the model's stability w.r.t. shifts in prediction
mechanisms ($Y|X$-shifts). Experimental results are consistent with our
intuition and validate the effectiveness of our algorithm. The code can be
found at https://github.com/LJSthu/SRM.
Related papers
- Mitigating Covariate Shift in Misspecified Regression with Applications
to Reinforcement Learning [39.02112341007981]
We study the effect of distribution shift in the presence of model misspecification.
We show that empirical risk minimization, or standard least squares regression, can result in undesirable misspecification amplification.
We develop a new algorithm that avoids this undesirable behavior, resulting in no misspecification amplification while still obtaining optimal statistical rates.
arXiv Detail & Related papers (2024-01-22T18:59:12Z) - Value-Distributional Model-Based Reinforcement Learning [59.758009422067]
Quantifying uncertainty about a policy's long-term performance is important to solve sequential decision-making tasks.
We study the problem from a model-based Bayesian reinforcement learning perspective.
We propose Epistemic Quantile-Regression (EQR), a model-based algorithm that learns a value distribution function.
arXiv Detail & Related papers (2023-08-12T14:59:19Z) - Learning Optimal Features via Partial Invariance [18.552839725370383]
Invariant Risk Minimization (IRM) is a popular framework that aims to learn robust models from multiple environments.
We show that IRM can over-constrain the predictor and to remedy this, we propose a relaxation via $textitpartial invariance$.
Several experiments, conducted both in linear settings as well as with deep neural networks on tasks over both language and image data, allow us to verify our conclusions.
arXiv Detail & Related papers (2023-01-28T02:48:14Z) - Domain-Adjusted Regression or: ERM May Already Learn Features Sufficient
for Out-of-Distribution Generalization [52.7137956951533]
We argue that devising simpler methods for learning predictors on existing features is a promising direction for future research.
We introduce Domain-Adjusted Regression (DARE), a convex objective for learning a linear predictor that is provably robust under a new model of distribution shift.
Under a natural model, we prove that the DARE solution is the minimax-optimal predictor for a constrained set of test distributions.
arXiv Detail & Related papers (2022-02-14T16:42:16Z) - Minimax Regret Optimization for Robust Machine Learning under
Distribution Shift [38.30154154957721]
We consider learning scenarios where the learned model is evaluated under an unknown test distribution.
We show that the DRO formulation does not guarantee uniformly small regret under distribution shift.
We propose an alternative method called Minimax Regret Optimization (MRO)
arXiv Detail & Related papers (2022-02-11T04:17:22Z) - Re-parameterizing VAEs for stability [1.90365714903665]
We propose a theoretical approach towards the training numerical stability of Variational AutoEncoders (VAE)
Our work is motivated by recent studies empowering VAEs to reach state of the art generative results on complex image datasets.
We show that by implementing small changes to the way we parameterize the Normal distributions on which they rely, VAEs can securely be trained.
arXiv Detail & Related papers (2021-06-25T16:19:09Z) - Stability and Generalization of the Decentralized Stochastic Gradient
Descent [17.63112147669365]
The stability and generalization of gradient-based methods provide valuable insights into the performance of machine learning models.
We first establish the stability and guarantees for the decentralized gradient descent.
Our results are built on a few common assumptions and that decentralization deteriorates of for the first time.
arXiv Detail & Related papers (2021-02-02T04:23:23Z) - Training Generative Adversarial Networks by Solving Ordinary
Differential Equations [54.23691425062034]
We study the continuous-time dynamics induced by GAN training.
From this perspective, we hypothesise that instabilities in training GANs arise from the integration error.
We experimentally verify that well-known ODE solvers (such as Runge-Kutta) can stabilise training.
arXiv Detail & Related papers (2020-10-28T15:23:49Z) - Learning Calibrated Uncertainties for Domain Shift: A Distributionally
Robust Learning Approach [150.8920602230832]
We propose a framework for learning calibrated uncertainties under domain shifts.
In particular, the density ratio estimation reflects the closeness of a target (test) sample to the source (training) distribution.
We show that our proposed method generates calibrated uncertainties that benefit downstream tasks.
arXiv Detail & Related papers (2020-10-08T02:10:54Z) - Balance-Subsampled Stable Prediction [55.13512328954456]
We propose a novel balance-subsampled stable prediction (BSSP) algorithm based on the theory of fractional factorial design.
A design-theoretic analysis shows that the proposed method can reduce the confounding effects among predictors induced by the distribution shift.
Numerical experiments on both synthetic and real-world data sets demonstrate that our BSSP algorithm significantly outperforms the baseline methods for stable prediction across unknown test data.
arXiv Detail & Related papers (2020-06-08T07:01:38Z) - Distributional Robustness and Regularization in Reinforcement Learning [62.23012916708608]
We introduce a new regularizer for empirical value functions and show that it lower bounds the Wasserstein distributionally robust value function.
It suggests using regularization as a practical tool for dealing with $textitexternal uncertainty$ in reinforcement learning.
arXiv Detail & Related papers (2020-03-05T19:56:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.