Stability Based Generalization Bounds for Exponential Family Langevin
Dynamics
- URL: http://arxiv.org/abs/2201.03064v1
- Date: Sun, 9 Jan 2022 18:15:22 GMT
- Title: Stability Based Generalization Bounds for Exponential Family Langevin
Dynamics
- Authors: Arindam Banerjee, Tiancong Chen, Xinyan Li and Yingxue Zhou
- Abstract summary: We study generalization bounds for noisy mini-batch iterative algorithms based on the notion of stability.
We introduce Exponential Family Langevin Dynamics(EFLD) which is a substantial generalization of SGLD and which allows exponential family noise to be used with gradient descent.
Third, we consider an important special case of EFLD: noisy sign-SGD, which extends sign-SGD using Bernoulli noise over -1,+1.
- Score: 21.26469220896393
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study generalization bounds for noisy stochastic mini-batch iterative
algorithms based on the notion of stability. Recent years have seen key
advances in data-dependent generalization bounds for noisy iterative learning
algorithms such as stochastic gradient Langevin dynamics (SGLD) based on
stability (Mou et al., 2018; Li et al., 2020) and information theoretic
approaches (Xu and Raginsky, 2017; Negrea et al., 2019; Steinke and
Zakynthinou, 2020; Haghifam et al., 2020). In this paper, we unify and
substantially generalize stability based generalization bounds and make three
technical advances. First, we bound the generalization error of general noisy
stochastic iterative algorithms (not necessarily gradient descent) in terms of
expected (not uniform) stability. The expected stability can in turn be bounded
by a Le Cam Style Divergence. Such bounds have a O(1/n) sample dependence
unlike many existing bounds with O(1/\sqrt{n}) dependence. Second, we introduce
Exponential Family Langevin Dynamics(EFLD) which is a substantial
generalization of SGLD and which allows exponential family noise to be used
with stochastic gradient descent (SGD). We establish data-dependent expected
stability based generalization bounds for general EFLD algorithms. Third, we
consider an important special case of EFLD: noisy sign-SGD, which extends
sign-SGD using Bernoulli noise over {-1,+1}. Generalization bounds for noisy
sign-SGD are implied by that of EFLD and we also establish optimization
guarantees for the algorithm. Further, we present empirical results on
benchmark datasets to illustrate that our bounds are non-vacuous and
quantitatively much sharper than existing bounds.
Related papers
- Breaking the Heavy-Tailed Noise Barrier in Stochastic Optimization Problems [56.86067111855056]
We consider clipped optimization problems with heavy-tailed noise with structured density.
We show that it is possible to get faster rates of convergence than $mathcalO(K-(alpha - 1)/alpha)$, when the gradients have finite moments of order.
We prove that the resulting estimates have negligible bias and controllable variance.
arXiv Detail & Related papers (2023-11-07T17:39:17Z) - The Implicit Regularization of Dynamical Stability in Stochastic
Gradient Descent [32.25490196411385]
We study the implicit regularization of gradient descent (SGD) through the lens of em dynamical stability
We analyze the generalization properties of two-layer ReLU networks and diagonal linear networks.
arXiv Detail & Related papers (2023-05-27T14:54:21Z) - Boosting the Confidence of Generalization for $L_2$-Stable Randomized
Learning Algorithms [41.082982732100696]
We show that a properly designed subbagging process leads to near-tight exponential generalization bounds over both data and algorithm.
We further derive generic results to improve high-probability generalization bounds for convex or non-tight problems with natural decaying learning rates.
arXiv Detail & Related papers (2022-06-08T12:14:01Z) - Utilising the CLT Structure in Stochastic Gradient based Sampling :
Improved Analysis and Faster Algorithms [14.174806471635403]
We consider approximations of sampling algorithms, such as Gradient Langevin Dynamics (SGLD) and the Random Batch Method (RBM) for Interacting Particle Dynamcs (IPD)
We observe that the noise introduced by the approximation is nearly Gaussian due to the Central Limit Theorem (CLT) while the driving Brownian motion is exactly Gaussian.
We harness this structure to absorb the approximation error inside the diffusion process, and obtain improved convergence guarantees for these algorithms.
arXiv Detail & Related papers (2022-06-08T10:17:40Z) - Clipped Stochastic Methods for Variational Inequalities with
Heavy-Tailed Noise [64.85879194013407]
We prove the first high-probability results with logarithmic dependence on the confidence level for methods for solving monotone and structured non-monotone VIPs.
Our results match the best-known ones in the light-tails case and are novel for structured non-monotone problems.
In addition, we numerically validate that the gradient noise of many practical formulations is heavy-tailed and show that clipping improves the performance of SEG/SGDA.
arXiv Detail & Related papers (2022-06-02T15:21:55Z) - Benign Underfitting of Stochastic Gradient Descent [72.38051710389732]
We study to what extent may gradient descent (SGD) be understood as a "conventional" learning rule that achieves generalization performance by obtaining a good fit training data.
We analyze the closely related with-replacement SGD, for which an analogous phenomenon does not occur and prove that its population risk does in fact converge at the optimal rate.
arXiv Detail & Related papers (2022-02-27T13:25:01Z) - Faster Convergence of Stochastic Gradient Langevin Dynamics for
Non-Log-Concave Sampling [110.88857917726276]
We provide a new convergence analysis of gradient Langevin dynamics (SGLD) for sampling from a class of distributions that can be non-log-concave.
At the core of our approach is a novel conductance analysis of SGLD using an auxiliary time-reversible Markov Chain.
arXiv Detail & Related papers (2020-10-19T15:23:18Z) - Fine-Grained Analysis of Stability and Generalization for Stochastic
Gradient Descent [55.85456985750134]
We introduce a new stability measure called on-average model stability, for which we develop novel bounds controlled by the risks of SGD iterates.
This yields generalization bounds depending on the behavior of the best model, and leads to the first-ever-known fast bounds in the low-noise setting.
To our best knowledge, this gives the firstever-known stability and generalization for SGD with even non-differentiable loss functions.
arXiv Detail & Related papers (2020-06-15T06:30:19Z) - Stability of Stochastic Gradient Descent on Nonsmooth Convex Losses [52.039438701530905]
We provide sharp upper and lower bounds for several forms of gradient descent (SGD) on arbitrary Lipschitz nonsmooth convex losses.
Our bounds allow us to derive a new algorithm for differentially private nonsmooth convex optimization with optimal excess population risk.
arXiv Detail & Related papers (2020-06-12T02:45:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.