Probabilistic Sequential Shrinking: A Best Arm Identification Algorithm
for Stochastic Bandits with Corruptions
- URL: http://arxiv.org/abs/2010.07904v4
- Date: Fri, 18 Jun 2021 10:36:07 GMT
- Title: Probabilistic Sequential Shrinking: A Best Arm Identification Algorithm
for Stochastic Bandits with Corruptions
- Authors: Zixin Zhong, Wang Chi Cheung, Vincent Y. F. Tan
- Abstract summary: We consider a best arm identification (BAI) problem for Sequential bandits with adversarial corruptions in the fixed-budget setting of T steps.
We design a novel randomized algorithm, Probabilistic Shrinking($u$) (PSS($u$)), which is agnostic to the amount of corruptions.
We show that when the CPS is sufficiently large, no algorithm can achieve a BAI probability tending to $1$ as $Trightarrow infty$.
- Score: 91.8283876874947
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We consider a best arm identification (BAI) problem for stochastic bandits
with adversarial corruptions in the fixed-budget setting of T steps. We design
a novel randomized algorithm, Probabilistic Sequential Shrinking($u$)
(PSS($u$)), which is agnostic to the amount of corruptions. When the amount of
corruptions per step (CPS) is below a threshold, PSS($u$) identifies the best
arm or item with probability tending to $1$ as $T\rightarrow \infty$.
Otherwise, the optimality gap of the identified item degrades gracefully with
the CPS.We argue that such a bifurcation is necessary. In PSS($u$), the
parameter $u$ serves to balance between the optimality gap and success
probability. The injection of randomization is shown to be essential to
mitigate the impact of corruptions. To demonstrate this, we design two attack
strategies that are applicable to any algorithm. We apply one of them to a
deterministic analogue of PSS($u$) known as Successive Halving (SH) by Karnin
et al. (2013). The attack strategy results in a high failure probability for
SH, but PSS($u$) remains robust. In the absence of corruptions, PSS($2$)'s
performance guarantee matches SH's. We show that when the CPS is sufficiently
large, no algorithm can achieve a BAI probability tending to $1$ as
$T\rightarrow \infty$. Numerical experiments corroborate our theoretical
findings.
Related papers
- Stochastic Bandits Robust to Adversarial Attacks [33.278131584647745]
This paper investigates multi-armed bandit algorithms that are robust to adversarial attacks.
We study two cases of this model, with or without the knowledge of an attack budget $C$.
We devise two types of algorithms with regret bounds having additive or multiplicative $C$ dependence terms.
arXiv Detail & Related papers (2024-08-16T17:41:35Z) - Replicability is Asymptotically Free in Multi-armed Bandits [45.729105054410745]
This work is motivated by the growing demand for reproducible machine learning.
In particular, we consider a replicable algorithm that ensures, with high probability, that the algorithm's sequence of actions is not affected by the randomness inherent in the dataset.
arXiv Detail & Related papers (2024-02-12T03:31:34Z) - Robust Lipschitz Bandits to Adversarial Corruptions [61.85150061213987]
Lipschitz bandit is a variant of bandits that deals with a continuous arm set defined on a metric space.
In this paper, we introduce a new problem of Lipschitz bandits in the presence of adversarial corruptions.
Our work presents the first line of robust Lipschitz bandit algorithms that can achieve sub-linear regret under both types of adversary.
arXiv Detail & Related papers (2023-05-29T18:16:59Z) - Nearly Optimal Algorithms for Linear Contextual Bandits with Adversarial
Corruptions [98.75618795470524]
We study the linear contextual bandit problem in the presence of adversarial corruption.
We propose a new algorithm based on the principle of optimism in the face of uncertainty.
arXiv Detail & Related papers (2022-05-13T17:58:58Z) - A Robust Phased Elimination Algorithm for Corruption-Tolerant Gaussian
Process Bandits [118.22458816174144]
We propose a novel robust elimination-type algorithm that runs in epochs, combines exploration with infrequent switching to select a small subset of actions, and plays each action for multiple time instants.
Our algorithm, GP Robust Phased Elimination (RGP-PE), successfully balances robustness to corruptions with exploration and exploitation.
We perform the first empirical study of robustness in the corrupted GP bandit setting, and show that our algorithm is robust against a variety of adversarial attacks.
arXiv Detail & Related papers (2022-02-03T21:19:36Z) - Linear Contextual Bandits with Adversarial Corruptions [91.38793800392108]
We study the linear contextual bandit problem in the presence of adversarial corruption.
We present a variance-aware algorithm that is adaptive to the level of adversarial contamination $C$.
arXiv Detail & Related papers (2021-10-25T02:53:24Z) - Stochastic Linear Bandits Robust to Adversarial Attacks [117.665995707568]
We provide two variants of a Robust Phased Elimination algorithm, one that knows $C$ and one that does not.
We show that both variants attain near-optimal regret in the non-corrupted case $C = 0$, while incurring additional additive terms respectively.
In a contextual setting, we show that a simple greedy algorithm is provably robust with a near-optimal additive regret term, despite performing no explicit exploration and not knowing $C$.
arXiv Detail & Related papers (2020-07-07T09:00:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.