AdaScale SGD: A User-Friendly Algorithm for Distributed Training
- URL: http://arxiv.org/abs/2007.05105v1
- Date: Thu, 9 Jul 2020 23:26:13 GMT
- Title: AdaScale SGD: A User-Friendly Algorithm for Distributed Training
- Authors: Tyler B. Johnson, Pulkit Agrawal, Haijie Gu, Carlos Guestrin
- Abstract summary: We propose AdaScale SGD, an algorithm that reliably adapts learning rates to large-batch training.
By continually adapting to the gradient's variance, AdaScale achieves speed-ups for a wide range of batch sizes.
This includes large-batch training with no model degradation for machine translation, image classification, object detection, and speech recognition tasks.
- Score: 29.430153773234363
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: When using large-batch training to speed up stochastic gradient descent,
learning rates must adapt to new batch sizes in order to maximize speed-ups and
preserve model quality. Re-tuning learning rates is resource intensive, while
fixed scaling rules often degrade model quality. We propose AdaScale SGD, an
algorithm that reliably adapts learning rates to large-batch training. By
continually adapting to the gradient's variance, AdaScale automatically
achieves speed-ups for a wide range of batch sizes. We formally describe this
quality with AdaScale's convergence bound, which maintains final objective
values, even as batch sizes grow large and the number of iterations decreases.
In empirical comparisons, AdaScale trains well beyond the batch size limits of
popular "linear learning rate scaling" rules. This includes large-batch
training with no model degradation for machine translation, image
classification, object detection, and speech recognition tasks. AdaScale's
qualitative behavior is similar to that of "warm-up" heuristics, but unlike
warm-up, this behavior emerges naturally from a principled mechanism. The
algorithm introduces negligible computational overhead and no new
hyperparameters, making AdaScale an attractive choice for large-scale training
in practice.
Related papers
- AdAdaGrad: Adaptive Batch Size Schemes for Adaptive Gradient Methods [17.043034606088234]
We introduce AdAdaGrad's scalar variant AdAdaGradNorm, which increase sizes during training.
We also perform image classification experiments, highlighting the merits of our proposed strategies.
arXiv Detail & Related papers (2024-02-17T07:49:50Z) - Winner-Take-All Column Row Sampling for Memory Efficient Adaptation of Language Model [89.8764435351222]
We propose a new family of unbiased estimators called WTA-CRS, for matrix production with reduced variance.
Our work provides both theoretical and experimental evidence that, in the context of tuning transformers, our proposed estimators exhibit lower variance compared to existing ones.
arXiv Detail & Related papers (2023-05-24T15:52:08Z) - Q-Ensemble for Offline RL: Don't Scale the Ensemble, Scale the Batch
Size [58.762959061522736]
We show that scaling mini-batch sizes with appropriate learning rate adjustments can speed up the training process by orders of magnitude.
We show that scaling the mini-batch size and naively adjusting the learning rate allows for (1) a reduced size of the Q-ensemble, (2) stronger penalization of out-of-distribution actions, and (3) improved convergence time.
arXiv Detail & Related papers (2022-11-20T21:48:25Z) - Scaling Forward Gradient With Local Losses [117.22685584919756]
Forward learning is a biologically plausible alternative to backprop for learning deep neural networks.
We show that it is possible to substantially reduce the variance of the forward gradient by applying perturbations to activations rather than weights.
Our approach matches backprop on MNIST and CIFAR-10 and significantly outperforms previously proposed backprop-free algorithms on ImageNet.
arXiv Detail & Related papers (2022-10-07T03:52:27Z) - Step-size Adaptation Using Exponentiated Gradient Updates [21.162404996362948]
We show that augmenting a given with an adaptive tuning method of the step-size greatly improves the performance.
We maintain a global step-size scale for the update as well as a gain factor for each coordinate.
We show that our approach can achieve compelling accuracy on standard models without using any specially tuned learning rate schedule.
arXiv Detail & Related papers (2022-01-31T23:17:08Z) - Automated Learning Rate Scheduler for Large-batch Training [24.20872850681828]
Large-batch training has been essential in leveraging large-scale datasets and models in deep learning.
It often requires a specially designed learning rate (LR) schedule to achieve a comparable level of performance as in smaller batch training.
We propose an automated LR scheduling algorithm which is effective for neural network training with a large batch size under the given epoch budget.
arXiv Detail & Related papers (2021-07-13T05:23:13Z) - Balancing Rates and Variance via Adaptive Batch-Size for Stochastic
Optimization Problems [120.21685755278509]
In this work, we seek to balance the fact that attenuating step-size is required for exact convergence with the fact that constant step-size learns faster in time up to an error.
Rather than fixing the minibatch the step-size at the outset, we propose to allow parameters to evolve adaptively.
arXiv Detail & Related papers (2020-07-02T16:02:02Z) - AdamP: Slowing Down the Slowdown for Momentum Optimizers on
Scale-invariant Weights [53.8489656709356]
Normalization techniques are a boon for modern deep learning.
It is often overlooked, however, that the additional introduction of momentum results in a rapid reduction in effective step sizes for scale-invariant weights.
In this paper, we verify that the widely-adopted combination of the two ingredients lead to the premature decay of effective step sizes and sub-optimal model performances.
arXiv Detail & Related papers (2020-06-15T08:35:15Z) - Extrapolation for Large-batch Training in Deep Learning [72.61259487233214]
We show that a host of variations can be covered in a unified framework that we propose.
We prove the convergence of this novel scheme and rigorously evaluate its empirical performance on ResNet, LSTM, and Transformer.
arXiv Detail & Related papers (2020-06-10T08:22:41Z) - Improving the convergence of SGD through adaptive batch sizes [0.1813006808606333]
Mini-batch gradient descent (SGD) and variants thereof approximate the objective function's gradient with a small number of training examples.
This work presents a method to adapt the batch size to the model's training loss.
arXiv Detail & Related papers (2019-10-18T01:45:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.