Dynamic Regret of Online Mirror Descent for Relatively Smooth Convex
Cost Functions
- URL: http://arxiv.org/abs/2202.12843v1
- Date: Fri, 25 Feb 2022 17:35:07 GMT
- Title: Dynamic Regret of Online Mirror Descent for Relatively Smooth Convex
Cost Functions
- Authors: Nima Eshraghi and Ben Liang
- Abstract summary: We show that it is possible to bound the dynamic regret, even when neither Lipschitz continuity nor uniform smoothness is present.
We then show that with an additional condition of relatively strong convexity, the dynamic regret can be bounded by the path length and gradient variation.
- Score: 30.412826613009518
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The performance of online convex optimization algorithms in a dynamic
environment is often expressed in terms of the dynamic regret, which measures
the decision maker's performance against a sequence of time-varying
comparators. In the analysis of the dynamic regret, prior works often assume
Lipschitz continuity or uniform smoothness of the cost functions. However,
there are many important cost functions in practice that do not satisfy these
conditions. In such cases, prior analyses are not applicable and fail to
guarantee the optimization performance. In this letter, we show that it is
possible to bound the dynamic regret, even when neither Lipschitz continuity
nor uniform smoothness is present. We adopt the notion of relative smoothness
with respect to some user-defined regularization function, which is a much
milder requirement on the cost functions. We first show that under relative
smoothness, the dynamic regret has an upper bound based on the path length and
functional variation. We then show that with an additional condition of
relatively strong convexity, the dynamic regret can be bounded by the path
length and gradient variation. These regret bounds provide performance
guarantees to a wide variety of online optimization problems that arise in
different application domains. Finally, we present numerical experiments that
demonstrate the advantage of adopting a regularization function under which the
cost functions are relatively smooth.
Related papers
- Stochastic Zeroth-Order Optimization under Strongly Convexity and Lipschitz Hessian: Minimax Sample Complexity [59.75300530380427]
We consider the problem of optimizing second-order smooth and strongly convex functions where the algorithm is only accessible to noisy evaluations of the objective function it queries.
We provide the first tight characterization for the rate of the minimax simple regret by developing matching upper and lower bounds.
arXiv Detail & Related papers (2024-06-28T02:56:22Z) - Contextual Continuum Bandits: Static Versus Dynamic Regret [70.71582850199871]
We study the contextual continuum bandits problem, where the learner sequentially receives a side information vector and has to choose an action in a convex set.
We demonstrate that any algorithm achieving a sub-linear static regret can be extended to achieve a sub-linear dynamic regret.
Inspired by the interior point method and employing self-concordant barriers, we propose an algorithm achieving a sub-linear dynamic regret.
arXiv Detail & Related papers (2024-06-09T10:12:08Z) - Improving Adaptive Online Learning Using Refined Discretization [44.646191058243645]
We study unconstrained Online Linear Optimization with Lipschitz losses.
Motivated by the pursuit of instance optimality, we propose a new algorithm.
Central to these results is a continuous time approach to online learning.
arXiv Detail & Related papers (2023-09-27T21:54:52Z) - Improved Dynamic Regret for Online Frank-Wolfe [54.690867216880356]
We investigate the dynamic regret of online Frank-Wolfe (OFW), which is an efficient projection-free algorithm for online convex optimization.
In this paper, we derive improved dynamic regret bounds for OFW by extending the fast convergence rates of FW from offline optimization to online optimization.
arXiv Detail & Related papers (2023-02-11T07:19:51Z) - Adaptivity and Non-stationarity: Problem-dependent Dynamic Regret for Online Convex Optimization [70.4342220499858]
We introduce novel online algorithms that can exploit smoothness and replace the dependence on $T$ in dynamic regret with problem-dependent quantities.
Our results are adaptive to the intrinsic difficulty of the problem, since the bounds are tighter than existing results for easy problems and safeguard the same rate in the worst case.
arXiv Detail & Related papers (2021-12-29T02:42:59Z) - Cumulative Regret Analysis of the Piyavskii--Shubert Algorithm and Its
Variants for Global Optimization [0.0]
We study the problem of global optimization, where we analyze the performance of the Piyavskii--Shubert algorithm and its variants.
We show that, our algorithm efficiently determines its queries; and achieves nearly minimax optimal (up to log factors) cumulative regret.
arXiv Detail & Related papers (2021-08-24T17:36:33Z) - Value Iteration in Continuous Actions, States and Time [99.00362538261972]
We propose a continuous fitted value iteration (cFVI) algorithm for continuous states and actions.
The optimal policy can be derived for non-linear control-affine dynamics.
Videos of the physical system are available at urlhttps://sites.google.com/view/value-iteration.
arXiv Detail & Related papers (2021-05-10T21:40:56Z) - A closer look at temporal variability in dynamic online learning [19.468067110814808]
This work focuses on the setting of dynamic regret in the context of online learning with full information.
By assuming that the sequence of loss functions does not vary much with time, we show that it is possible to incur improved regret bounds compared to existing results.
arXiv Detail & Related papers (2021-02-15T16:50:16Z) - Dynamic Regret of Convex and Smooth Functions [93.71361250701075]
We investigate online convex optimization in non-stationary environments.
We choose the dynamic regret as the performance measure.
We show that it is possible to further enhance the dynamic regret by exploiting the smoothness condition.
arXiv Detail & Related papers (2020-07-07T14:10:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.