Delay-adaptive step-sizes for asynchronous learning
- URL: http://arxiv.org/abs/2202.08550v1
- Date: Thu, 17 Feb 2022 09:51:22 GMT
- Title: Delay-adaptive step-sizes for asynchronous learning
- Authors: Xuyang Wu, Sindri Magnusson, Hamid Reza Feyzmahdavian and Mikael
Johansson
- Abstract summary: We show that it is possible to use learning rates that depend on the actual time-varying delays in the system.
For each of these methods, we demonstrate how delays can be measured on-line, present delay-adaptive step-size policies, and illustrate their theoretical and practical advantages over the state-of-the-art.
- Score: 8.272788656521415
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In scalable machine learning systems, model training is often parallelized
over multiple nodes that run without tight synchronization. Most analysis
results for the related asynchronous algorithms use an upper bound on the
information delays in the system to determine learning rates. Not only are such
bounds hard to obtain in advance, but they also result in unnecessarily slow
convergence. In this paper, we show that it is possible to use learning rates
that depend on the actual time-varying delays in the system. We develop general
convergence results for delay-adaptive asynchronous iterations and specialize
these to proximal incremental gradient descent and block-coordinate descent
algorithms. For each of these methods, we demonstrate how delays can be
measured on-line, present delay-adaptive step-size policies, and illustrate
their theoretical and practical advantages over the state-of-the-art.
Related papers
Err
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.