Non-stationary Projection-free Online Learning with Dynamic and Adaptive
Regret Guarantees
- URL: http://arxiv.org/abs/2305.11726v1
- Date: Fri, 19 May 2023 15:02:10 GMT
- Title: Non-stationary Projection-free Online Learning with Dynamic and Adaptive
Regret Guarantees
- Authors: Yibo Wang, Wenhao Yang, Wei Jiang, Shiyin Lu, Bing Wang, Haihong Tang,
Yuanyu Wan, Lijun Zhang
- Abstract summary: We investigate non-stationary projection-free online learning, and choose dynamic regret and adaptive regret to measure the performance.
Our results are the first general-case dynamic regret bounds for projection-free online learning, and can recover the existing $mathcalO(T3/4)$ static regret.
We propose a projection-free method to attain an $tildemathcalO(tau3/4)$ adaptive regret bound for any interval with length $tau.
- Score: 36.746745619968024
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Projection-free online learning has drawn increasing interest due to its
efficiency in solving high-dimensional problems with complicated constraints.
However, most existing projection-free online methods focus on minimizing the
static regret, which unfortunately fails to capture the challenge of changing
environments. In this paper, we investigate non-stationary projection-free
online learning, and choose dynamic regret and adaptive regret to measure the
performance. Specifically, we first provide a novel dynamic regret analysis for
an existing projection-free method named $\text{BOGD}_\text{IP}$, and establish
an $\mathcal{O}(T^{3/4}(1+P_T))$ dynamic regret bound, where $P_T$ denotes the
path-length of the comparator sequence. Then, we improve the upper bound to
$\mathcal{O}(T^{3/4}(1+P_T)^{1/4})$ by running multiple $\text{BOGD}_\text{IP}$
algorithms with different step sizes in parallel, and tracking the best one on
the fly. Our results are the first general-case dynamic regret bounds for
projection-free online learning, and can recover the existing
$\mathcal{O}(T^{3/4})$ static regret by setting $P_T = 0$. Furthermore, we
propose a projection-free method to attain an $\tilde{\mathcal{O}}(\tau^{3/4})$
adaptive regret bound for any interval with length $\tau$, which nearly matches
the static regret over that interval. The essential idea is to maintain a set
of $\text{BOGD}_\text{IP}$ algorithms dynamically, and combine them by a meta
algorithm. Moreover, we demonstrate that it is also equipped with an
$\mathcal{O}(T^{3/4}(1+P_T)^{1/4})$ dynamic regret bound. Finally, empirical
studies verify our theoretical findings.
Related papers
- Online Convex Optimization with a Separation Oracle [10.225358400539719]
We introduce a new projection-free algorithm for Online Convex Optimization (OCO) with a state-of-the-art regret guarantee.
Our algorithm achieves a regret bound of $widetildeO(sqrtdT + kappa d)$, while requiring only $widetildeO(1) calls to a separation oracle per round.
arXiv Detail & Related papers (2024-10-03T13:35:08Z) - An Equivalence Between Static and Dynamic Regret Minimization [10.812831455376218]
We show that for linear losses, dynamic regret minimization is equivalent to static regret minimization in an extended decision space.
We provide an algorithm guaranteeing dynamic regret of the form $R_T(u_1,dots,u_T)le tilde.
arXiv Detail & Related papers (2024-06-03T17:54:58Z) - Improved Projection-free Online Continuous Submodular Maximization [35.324719857218014]
We investigate the problem of online learning with monotone and continuous DR-submodular reward functions.
Previous studies have proposed an efficient projection-free algorithm called Mono-Frank-Wolfe (Mono-FW) using $O(T)$ gradient evaluations.
We propose an improved projection-free algorithm, namely POBGA, which reduces the regret bound to $O(T3/4)$ while keeping the same computational complexity.
arXiv Detail & Related papers (2023-05-29T02:54:31Z) - Improved Dynamic Regret for Online Frank-Wolfe [54.690867216880356]
We investigate the dynamic regret of online Frank-Wolfe (OFW), which is an efficient projection-free algorithm for online convex optimization.
In this paper, we derive improved dynamic regret bounds for OFW by extending the fast convergence rates of FW from offline optimization to online optimization.
arXiv Detail & Related papers (2023-02-11T07:19:51Z) - Logarithmic Regret from Sublinear Hints [76.87432703516942]
We show that an algorithm can obtain $O(log T)$ regret with just $O(sqrtT)$ hints under a natural query model.
We also show that $o(sqrtT)$ hints cannot guarantee better than $Omega(sqrtT)$ regret.
arXiv Detail & Related papers (2021-11-09T16:50:18Z) - Online Convex Optimization with Continuous Switching Constraint [78.25064451417082]
We introduce the problem of online convex optimization with continuous switching constraint.
We show that, for strongly convex functions, the regret bound can be improved to $O(log T)$ for $S=Omega(log T)$, and $O(minT/exp(S)+S,T)$ for $S=O(log T)$.
arXiv Detail & Related papers (2021-03-21T11:43:35Z) - Optimal Regret Algorithm for Pseudo-1d Bandit Convex Optimization [51.23789922123412]
We study online learning with bandit feedback (i.e. learner has access to only zeroth-order oracle) where cost/reward functions admit a "pseudo-1d" structure.
We show a lower bound of $min(sqrtdT, T3/4)$ for the regret of any algorithm, where $T$ is the number of rounds.
We propose a new algorithm sbcalg that combines randomized online gradient descent with a kernelized exponential weights method to exploit the pseudo-1d structure effectively.
arXiv Detail & Related papers (2021-02-15T08:16:51Z) - Revisiting Smoothed Online Learning [70.09792747315323]
We investigate the problem of smoothed online learning, in which the online learner suffers both a hitting cost and a switching cost.
To bound the competitive ratio, we assume the hitting cost is known to the learner in each round, and investigate the greedy algorithm which simply minimizes the weighted sum of the hitting cost and the switching cost.
arXiv Detail & Related papers (2021-02-13T14:15:55Z) - Improved Analysis for Dynamic Regret of Strongly Convex and Smooth
Functions [24.445848532264307]
Original analysis shows that the dynamic regret of OMGD is at most $mathcalO(minmathcalP_T,mathcalS_T)$, where $mathcalP_T$ and $mathcalS_T$ are path-length and squared path-length.
We demonstrate that by an improved analysis, the dynamic regret of OMGD can be improved to $mathcalO(minmathcalP_T,mathcalS_T,mathcal
arXiv Detail & Related papers (2020-06-10T15:01:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.