Unconstrained Online Optimization: Dynamic Regret Analysis of Strongly
Convex and Smooth Problems
- URL: http://arxiv.org/abs/2006.03912v2
- Date: Fri, 14 Aug 2020 17:28:53 GMT
- Title: Unconstrained Online Optimization: Dynamic Regret Analysis of Strongly
Convex and Smooth Problems
- Authors: Ting-Jui Chang, Shahin Shahrampour
- Abstract summary: We propose an optimistic Newton method for the case when the first and second order information of the function sequence is predictable.
By using multiple gradients for OGD, we can achieve an upper bound of $O(minC*_2,T,V_T)$ on regret.
- Score: 14.924672048447334
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The regret bound of dynamic online learning algorithms is often expressed in
terms of the variation in the function sequence ($V_T$) and/or the path-length
of the minimizer sequence after $T$ rounds. For strongly convex and smooth
functions, , Zhang et al. establish the squared path-length of the minimizer
sequence ($C^*_{2,T}$) as a lower bound on regret. They also show that online
gradient descent (OGD) achieves this lower bound using multiple gradient
queries per round. In this paper, we focus on unconstrained online
optimization. We first show that a preconditioned variant of OGD achieves
$O(C^*_{2,T})$ with one gradient query per round. We then propose online
optimistic Newton (OON) method for the case when the first and second order
information of the function sequence is predictable. The regret bound of OON is
captured via the quartic path-length of the minimizer sequence ($C^*_{4,T}$),
which can be much smaller than $C^*_{2,T}$. We finally show that by using
multiple gradients for OGD, we can achieve an upper bound of
$O(\min\{C^*_{2,T},V_T\})$ on regret.
Related papers
- Non-stationary Online Convex Optimization with Arbitrary Delays [50.46856739179311]
This paper investigates the delayed online convex optimization (OCO) in non-stationary environments.
We first propose a simple algorithm, namely DOGD, which performs a gradient descent step for each delayed gradient according to their arrival order.
We develop an improved algorithm, which reduces those dynamic regret bounds achieved by DOGD to $O(sqrtbardT(P_T+1))$.
arXiv Detail & Related papers (2023-05-20T07:54:07Z) - Non-stationary Projection-free Online Learning with Dynamic and Adaptive
Regret Guarantees [36.746745619968024]
We investigate non-stationary projection-free online learning, and choose dynamic regret and adaptive regret to measure the performance.
Our results are the first general-case dynamic regret bounds for projection-free online learning, and can recover the existing $mathcalO(T3/4)$ static regret.
We propose a projection-free method to attain an $tildemathcalO(tau3/4)$ adaptive regret bound for any interval with length $tau.
arXiv Detail & Related papers (2023-05-19T15:02:10Z) - Improved Dynamic Regret for Online Frank-Wolfe [54.690867216880356]
We investigate the dynamic regret of online Frank-Wolfe (OFW), which is an efficient projection-free algorithm for online convex optimization.
In this paper, we derive improved dynamic regret bounds for OFW by extending the fast convergence rates of FW from offline optimization to online optimization.
arXiv Detail & Related papers (2023-02-11T07:19:51Z) - A first-order primal-dual method with adaptivity to local smoothness [64.62056765216386]
We consider the problem of finding a saddle point for the convex-concave objective $min_x max_y f(x) + langle Ax, yrangle - g*(y)$, where $f$ is a convex function with locally Lipschitz gradient and $g$ is convex and possibly non-smooth.
We propose an adaptive version of the Condat-Vu algorithm, which alternates between primal gradient steps and dual steps.
arXiv Detail & Related papers (2021-10-28T14:19:30Z) - Regret Bounds for Stochastic Shortest Path Problems with Linear Function
Approximation [29.549633678730054]
We propose two algorithms for episodic shortest path problems with linear function approximation.
The first is computationally expensive but provably obtains $tildeO (sqrtB_star3 d3 K/cmin )$ regret, where $B_star$ is a upper bound on the optimal cost-to-go function.
The second is computationally efficient in practice, and we conjecture that it obtains the same regret bound.
arXiv Detail & Related papers (2021-05-04T16:05:08Z) - Private Stochastic Convex Optimization: Optimal Rates in $\ell_1$
Geometry [69.24618367447101]
Up to logarithmic factors the optimal excess population loss of any $(varepsilon,delta)$-differently private is $sqrtlog(d)/n + sqrtd/varepsilon n.$
We show that when the loss functions satisfy additional smoothness assumptions, the excess loss is upper bounded (up to logarithmic factors) by $sqrtlog(d)/n + (log(d)/varepsilon n)2/3.
arXiv Detail & Related papers (2021-03-02T06:53:44Z) - Optimal Regret Algorithm for Pseudo-1d Bandit Convex Optimization [51.23789922123412]
We study online learning with bandit feedback (i.e. learner has access to only zeroth-order oracle) where cost/reward functions admit a "pseudo-1d" structure.
We show a lower bound of $min(sqrtdT, T3/4)$ for the regret of any algorithm, where $T$ is the number of rounds.
We propose a new algorithm sbcalg that combines randomized online gradient descent with a kernelized exponential weights method to exploit the pseudo-1d structure effectively.
arXiv Detail & Related papers (2021-02-15T08:16:51Z) - Gradient Free Minimax Optimization: Variance Reduction and Faster
Convergence [120.9336529957224]
In this paper, we denote the non-strongly setting on the magnitude of a gradient-free minimax optimization problem.
We show that a novel zeroth-order variance reduced descent algorithm achieves the best known query complexity.
arXiv Detail & Related papers (2020-06-16T17:55:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.