Asymptotic behaviour of learning rates in Armijo's condition
- URL: http://arxiv.org/abs/2007.03618v1
- Date: Tue, 7 Jul 2020 16:49:25 GMT
- Title: Asymptotic behaviour of learning rates in Armijo's condition
- Authors: Tuyen Trung Truong, Tuan Hang Nguyen
- Abstract summary: We show that if $x_n$ converges to a non-degenerate critical point, then $delta _n$ must be bounded.
This complements the first author's results on Unbounded Backtracking GD.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Fix a constant $0<\alpha <1$. For a $C^1$ function $f:\mathbb{R}^k\rightarrow
\mathbb{R}$, a point $x$ and a positive number $\delta >0$, we say that
Armijo's condition is satisfied if $f(x-\delta \nabla f(x))-f(x)\leq -\alpha
\delta ||\nabla f(x)||^2$. It is a basis for the well known Backtracking
Gradient Descent (Backtracking GD) algorithm.
Consider a sequence $\{x_n\}$ defined by $x_{n+1}=x_n-\delta _n\nabla
f(x_n)$, for positive numbers $\delta _n$ for which Armijo's condition is
satisfied. We show that if $\{x_n\}$ converges to a non-degenerate critical
point, then $\{\delta _n\}$ must be bounded. Moreover this boundedness can be
quantified in terms of the norms of the Hessian $\nabla ^2f$ and its inverse at
the limit point. This complements the first author's results on Unbounded
Backtracking GD, and shows that in case of convergence to a non-degenerate
critical point the behaviour of Unbounded Backtracking GD is not too different
from that of usual Backtracking GD. On the other hand, in case of convergence
to a degenerate critical point the behaviours can be very much different. We
run some experiments to illustrate that both scenrios can really happen.
In another part of the paper, we argue that Backtracking GD has the correct
unit (according to a definition by Zeiler in his Adadelta's paper). The main
point is that since learning rate in Backtracking GD is bound by Armijo's
condition, it is not unitless.
Related papers
- Efficient Continual Finite-Sum Minimization [52.5238287567572]
We propose a key twist into the finite-sum minimization, dubbed as continual finite-sum minimization.
Our approach significantly improves upon the $mathcalO(n/epsilon)$ FOs that $mathrmStochasticGradientDescent$ requires.
We also prove that there is no natural first-order method with $mathcalOleft(n/epsilonalpharight)$ complexity gradient for $alpha 1/4$, establishing that the first-order complexity of our method is nearly tight.
arXiv Detail & Related papers (2024-06-07T08:26:31Z) - Convergence of Gradient Descent with Small Initialization for
Unregularized Matrix Completion [21.846732043706318]
We show that the vanilla gradient descent provably converges to the ground truth $rmXstar$ without requiring any explicit regularization.
Surprisingly, neither the convergence rate nor the final accuracy depends on the over- parameterized search rank $r'$, and they are only governed by the true rank $r$.
arXiv Detail & Related papers (2024-02-09T19:39:23Z) - On the $O(\frac{\sqrt{d}}{T^{1/4}})$ Convergence Rate of RMSProp and Its Momentum Extension Measured by $\ell_1$ Norm [59.65871549878937]
This paper considers the RMSProp and its momentum extension and establishes the convergence rate of $frac1Tsum_k=1T.
Our convergence rate matches the lower bound with respect to all the coefficients except the dimension $d$.
Our convergence rate can be considered to be analogous to the $frac1Tsum_k=1T.
arXiv Detail & Related papers (2024-02-01T07:21:32Z) - A Unified Framework for Uniform Signal Recovery in Nonlinear Generative
Compressed Sensing [68.80803866919123]
Under nonlinear measurements, most prior results are non-uniform, i.e., they hold with high probability for a fixed $mathbfx*$ rather than for all $mathbfx*$ simultaneously.
Our framework accommodates GCS with 1-bit/uniformly quantized observations and single index models as canonical examples.
We also develop a concentration inequality that produces tighter bounds for product processes whose index sets have low metric entropy.
arXiv Detail & Related papers (2023-09-25T17:54:19Z) - Optimal Sketching Bounds for Sparse Linear Regression [116.30196615349226]
We study oblivious sketching for $k$-sparse linear regression under various loss functions such as an $ell_p$ norm, or from a broad class of hinge-like loss functions.
We show that for sparse $ell$vareps regression, there is an oblivious distribution over sketches with $Theta(klog(d/k)/varepsilon2)$ rows, which is tight up to a constant factor.
We also show that sketching $O(mu2 klog(mu n d/varepsilon)/var
arXiv Detail & Related papers (2023-04-05T07:24:19Z) - Spiked Covariance Estimation from Modulo-Reduced Measurements [14.569322713960494]
We develop and analyze an algorithm that, for most directions $bfu$ and $nu=mathrmpoly(k)$, estimates $bfu$ to high accuracy using $n=mathrmpoly(k)$ measurements.
Numerical experiments show that the developed algorithm performs well even in a non-asymptotic setting.
arXiv Detail & Related papers (2021-10-04T02:10:47Z) - Optimal Regret Algorithm for Pseudo-1d Bandit Convex Optimization [51.23789922123412]
We study online learning with bandit feedback (i.e. learner has access to only zeroth-order oracle) where cost/reward functions admit a "pseudo-1d" structure.
We show a lower bound of $min(sqrtdT, T3/4)$ for the regret of any algorithm, where $T$ is the number of rounds.
We propose a new algorithm sbcalg that combines randomized online gradient descent with a kernelized exponential weights method to exploit the pseudo-1d structure effectively.
arXiv Detail & Related papers (2021-02-15T08:16:51Z) - Coordinate-wise Armijo's condition: General case [0.0]
We prove convergent results for some functions such as $f(x,y)=f(x,y)+g(y)$.
We then analyse and present experimental results for some functions such as $f(x,y)=f(x,y)+g(y)$.
arXiv Detail & Related papers (2020-03-11T12:17:05Z) - Some convergent results for Backtracking Gradient Descent method on
Banach spaces [0.0]
bf Theorem. Let $X$ be a Banach space and $f:Xrightarrow mathbbR$ be a $C2$ function.
Let $mathcalC$ be the set of critical points of $f$.
arXiv Detail & Related papers (2020-01-16T12:49:42Z) - Backtracking Gradient Descent allowing unbounded learning rates [0.0]
In unconstrained optimisation on an Euclidean space, to prove convergence in Gradient Descent processes (GD) $x_n+1=x_n-delta _n nabla f(x_n)$ it usually is required that the learning rates $delta _n$'s are bounded.
In this paper, we allow the learning rates $delta _n$ to be unbounded.
It will be shown that this growth rate of $h$ is best possible if one wants convergence of the sequence $x_n$.
arXiv Detail & Related papers (2020-01-07T12:52:00Z) - On Gradient Descent Ascent for Nonconvex-Concave Minimax Problems [86.92205445270427]
We consider non-con minimax problems, $min_mathbfx max_mathhidoty f(mathbfdoty)$ efficiently.
arXiv Detail & Related papers (2019-06-02T03:03:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.