Machine Learning's Dropout Training is Distributionally Robust Optimal
- URL: http://arxiv.org/abs/2009.06111v2
- Date: Wed, 14 Apr 2021 05:29:05 GMT
- Title: Machine Learning's Dropout Training is Distributionally Robust Optimal
- Authors: Jose Blanchet and Yang Kang and Jose Luis Montiel Olea and Viet Anh
Nguyen and Xuhui Zhang
- Abstract summary: This paper shows that dropout training in Generalized Linear Models provides out-of-sample expected loss guarantees.
It also provides a novel, parallelizable, Unbiased Multi-Level Monte Carlo algorithm to speed-up the implementation of dropout training.
- Score: 10.937094979510212
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper shows that dropout training in Generalized Linear Models is the
minimax solution of a two-player, zero-sum game where an adversarial nature
corrupts a statistician's covariates using a multiplicative nonparametric
errors-in-variables model. In this game, nature's least favorable distribution
is dropout noise, where nature independently deletes entries of the covariate
vector with some fixed probability $\delta$. This result implies that dropout
training indeed provides out-of-sample expected loss guarantees for
distributions that arise from multiplicative perturbations of in-sample data.
In addition to the decision-theoretic analysis, the paper makes two more
contributions. First, there is a concrete recommendation on how to select the
tuning parameter $\delta$ to guarantee that, as the sample size grows large,
the in-sample loss after dropout training exceeds the true population loss with
some pre-specified probability. Second, the paper provides a novel,
parallelizable, Unbiased Multi-Level Monte Carlo algorithm to speed-up the
implementation of dropout training. Our algorithm has a much smaller
computational cost compared to the naive implementation of dropout, provided
the number of data points is much smaller than the dimension of the covariate
vector.
Related papers
- Robust Time Series Forecasting with Non-Heavy-Tailed Gaussian Loss-Weighted Sampler [1.8816077341295625]
Recent resampling methods aim to increase training efficiency by reweighting samples based on their running losses.
We introduce a novel approach: a Gaussian loss-weighted sampler that multiplies their running losses with a Gaussian distribution weight.
It reduces the probability of selecting samples with very low or very high losses while favoring those close to average losses.
arXiv Detail & Related papers (2024-06-19T22:28:18Z) - DistPred: A Distribution-Free Probabilistic Inference Method for Regression and Forecasting [14.390842560217743]
We propose a novel approach called DistPred for regression and forecasting tasks.
We transform proper scoring rules that measure the discrepancy between the predicted distribution and the target distribution into a differentiable discrete form.
This allows the model to sample numerous samples in a single forward pass to estimate the potential distribution of the response variable.
arXiv Detail & Related papers (2024-06-17T10:33:00Z) - Rejection via Learning Density Ratios [50.91522897152437]
Classification with rejection emerges as a learning paradigm which allows models to abstain from making predictions.
We propose a different distributional perspective, where we seek to find an idealized data distribution which maximizes a pretrained model's performance.
Our framework is tested empirically over clean and noisy datasets.
arXiv Detail & Related papers (2024-05-29T01:32:17Z) - Probabilistic Contrastive Learning for Long-Tailed Visual Recognition [78.70453964041718]
Longtailed distributions frequently emerge in real-world data, where a large number of minority categories contain a limited number of samples.
Recent investigations have revealed that supervised contrastive learning exhibits promising potential in alleviating the data imbalance.
We propose a novel probabilistic contrastive (ProCo) learning algorithm that estimates the data distribution of the samples from each class in the feature space.
arXiv Detail & Related papers (2024-03-11T13:44:49Z) - Bias Mimicking: A Simple Sampling Approach for Bias Mitigation [57.17709477668213]
We introduce a new class-conditioned sampling method: Bias Mimicking.
Bias Mimicking improves underrepresented groups' accuracy of sampling methods by 3% over four benchmarks.
arXiv Detail & Related papers (2022-09-30T17:33:00Z) - Learning from a Biased Sample [3.546358664345473]
We propose a method for learning a decision rule that minimizes the worst-case risk incurred under a family of test distributions.
We empirically validate our proposed method in a case study on prediction of mental health scores from health survey data.
arXiv Detail & Related papers (2022-09-05T04:19:16Z) - Unrolling Particles: Unsupervised Learning of Sampling Distributions [102.72972137287728]
Particle filtering is used to compute good nonlinear estimates of complex systems.
We show in simulations that the resulting particle filter yields good estimates in a wide range of scenarios.
arXiv Detail & Related papers (2021-10-06T16:58:34Z) - Examining and Combating Spurious Features under Distribution Shift [94.31956965507085]
We define and analyze robust and spurious representations using the information-theoretic concept of minimal sufficient statistics.
We prove that even when there is only bias of the input distribution, models can still pick up spurious features from their training data.
Inspired by our analysis, we demonstrate that group DRO can fail when groups do not directly account for various spurious correlations.
arXiv Detail & Related papers (2021-06-14T05:39:09Z) - Distributional Reinforcement Learning via Moment Matching [54.16108052278444]
We formulate a method that learns a finite set of statistics from each return distribution via neural networks.
Our method can be interpreted as implicitly matching all orders of moments between a return distribution and its Bellman target.
Experiments on the suite of Atari games show that our method outperforms the standard distributional RL baselines.
arXiv Detail & Related papers (2020-07-24T05:18:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.