Byzantine-Robust Federated Linear Bandits
- URL: http://arxiv.org/abs/2204.01155v1
- Date: Sun, 3 Apr 2022 20:22:26 GMT
- Title: Byzantine-Robust Federated Linear Bandits
- Authors: Ali Jadbabaie, Haochuan Li, Jian Qian, Yi Tian
- Abstract summary: We study a linear bandit optimization problem in a federated setting where a large collection of distributed agents collaboratively learn a common linear bandit model.
Standard federated learning algorithms are vulnerable to Byzantine attacks on even a small fraction of agents.
We show that our proposed algorithm is robust to Byzantine attacks on fewer than half of agents and achieves a sublinear $tildemathcalO(T3/4)$ regret with $mathcalO(sqrtT)$ steps of communication.
- Score: 27.77095339417368
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we study a linear bandit optimization problem in a federated
setting where a large collection of distributed agents collaboratively learn a
common linear bandit model. Standard federated learning algorithms applied to
this setting are vulnerable to Byzantine attacks on even a small fraction of
agents. We propose a novel algorithm with a robust aggregation oracle that
utilizes the geometric median. We prove that our proposed algorithm is robust
to Byzantine attacks on fewer than half of agents and achieves a sublinear
$\tilde{\mathcal{O}}({T^{3/4}})$ regret with $\mathcal{O}(\sqrt{T})$ steps of
communication in $T$ steps. Moreover, we make our algorithm differentially
private via a tree-based mechanism. Finally, if the level of corruption is
known to be small, we show that using the geometric median of mean oracle for
robust aggregation further improves the regret bound.
Related papers
- A Sample Efficient Alternating Minimization-based Algorithm For Robust Phase Retrieval [56.67706781191521]
In this work, we present a robust phase retrieval problem where the task is to recover an unknown signal.
Our proposed oracle avoids the need for computationally spectral descent, using a simple gradient step and outliers.
arXiv Detail & Related papers (2024-09-07T06:37:23Z) - Cooperative Thresholded Lasso for Sparse Linear Bandit [6.52540785559241]
We present a novel approach to address the multi-agent sparse contextual linear bandit problem.
It is first algorithm that tackles row-wise distributed data in sparse linear bandits.
It is widely applicable to high-dimensional multi-agent problems where efficient feature extraction is critical for minimizing regret.
arXiv Detail & Related papers (2023-05-30T16:05:44Z) - Differentially-Private Hierarchical Clustering with Provable
Approximation Guarantees [79.59010418610625]
We study differentially private approximation algorithms for hierarchical clustering.
We show strong lower bounds for the problem: that any $epsilon$-DP algorithm must exhibit $O(|V|2/ epsilon)$-additive error for an input dataset.
We propose a private $1+o(1)$ approximation algorithm which also recovers the blocks exactly.
arXiv Detail & Related papers (2023-01-31T19:14:30Z) - Corruption-Robust Algorithms with Uncertainty Weighting for Nonlinear
Contextual Bandits and Markov Decision Processes [59.61248760134937]
We propose an efficient algorithm to achieve a regret of $tildeO(sqrtT+zeta)$.
The proposed algorithm relies on the recently developed uncertainty-weighted least-squares regression from linear contextual bandit.
We generalize our algorithm to the episodic MDP setting and first achieve an additive dependence on the corruption level $zeta$.
arXiv Detail & Related papers (2022-12-12T15:04:56Z) - Collaborative Linear Bandits with Adversarial Agents: Near-Optimal
Regret Bounds [31.5504566292927]
We consider a linear bandit problem involving $M$ agents that can collaborate via a central server to minimize regret.
A fraction $alpha$ of these agents are adversarial and can act arbitrarily, leading to the following tension.
We design new algorithms that balance the exploration-exploitation trade-off via carefully constructed robust confidence intervals.
arXiv Detail & Related papers (2022-06-06T18:16:34Z) - Escaping Saddle Points in Distributed Newton's Method with Communication
efficiency and Byzantine Resilience [49.379254729853756]
We consider the problem of optimizing a non-regularized loss function (with saddle points) in a distributed framework in the presence of Byzantine machines.
We robustify the cubic-regularized Newton algorithm such that it avoids the saddle points and the fake local minimas efficiently.
We obtain theoretical guarantees for our proposed scheme under several approximate settings including (sub-sampled) and Hessians.
arXiv Detail & Related papers (2021-03-17T03:53:58Z) - Upper Confidence Bounds for Combining Stochastic Bandits [52.10197476419621]
We provide a simple method to combine bandit algorithms.
Our approach is based on a "meta-UCB" procedure that treats each of $N$ individual bandit algorithms as arms in a higher-level $N$-armed bandit problem.
arXiv Detail & Related papers (2020-12-24T05:36:29Z) - Decentralized Multi-Agent Linear Bandits with Safety Constraints [31.67685495996986]
We study decentralized linear bandits, where a network of $N$ agents acts cooperatively to solve a linear bandit-optimization problem.
We propose DLUCB: a fully decentralized algorithm that minimizes the cumulative regret over the entire network.
We show that our ideas extend naturally to the emerging, albeit more challenging, setting of safe bandits.
arXiv Detail & Related papers (2020-12-01T07:33:00Z) - Consistent $k$-Median: Simpler, Better and Robust [20.692372082600972]
We show that a simple local-search based online algorithm can give a bicriteria constant approximation for the problem with $O(k2 log2 (nD))$ swaps of medians (recourse) in total.
When restricted to the problem without outliers, our algorithm is simpler, deterministic and gives better approximation ratio and recourse.
arXiv Detail & Related papers (2020-08-13T20:24:28Z) - An Empirical Process Approach to the Union Bound: Practical Algorithms
for Combinatorial and Linear Bandits [34.06611065493047]
This paper proposes near-optimal algorithms for the pure-exploration linear bandit problem in the fixed confidence and fixed budget settings.
We provide an algorithm whose sample complexity scales with the geometry of the instance and avoids an explicit union bound over the number of arms.
We also propose the first algorithm for linear bandits in the the fixed budget setting.
arXiv Detail & Related papers (2020-06-21T00:56:33Z) - Stochastic Bandits with Linear Constraints [69.757694218456]
We study a constrained contextual linear bandit setting, where the goal of the agent is to produce a sequence of policies.
We propose an upper-confidence bound algorithm for this problem, called optimistic pessimistic linear bandit (OPLB)
arXiv Detail & Related papers (2020-06-17T22:32:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.