Distributed Online Learning for Joint Regret with Communication
Constraints
- URL: http://arxiv.org/abs/2102.07521v1
- Date: Mon, 15 Feb 2021 12:48:33 GMT
- Title: Distributed Online Learning for Joint Regret with Communication
Constraints
- Authors: Dirk van der Hoeven, H\'edi Hadiji, Tim van Erven
- Abstract summary: We consider a distributed online learning setting for joint regret with communication constraints.
A subset of all the agents may then communicate a $b$-bit message to their neighbors in a graph.
We exploit the comparator-adaptive property of our algorithm to learn the best partition from a set of candidate partitions.
- Score: 17.080853582489073
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper we consider a distributed online learning
setting for joint regret with communication constraints. This is a
multi-agent setting in which in each round $t$ an adversary activates an agent,
which has to issue a prediction. A subset of all the agents may then
communicate a $b$-bit message to their neighbors in a graph. All agents
cooperate to control the joint regret, which is the sum of the losses of the
agents minus the losses evaluated at the best fixed common comparator
parameters $\pmb{u}$. We provide a comparator-adaptive algorithm for this
setting, which means that the joint regret scales with the norm of the
comparator $\|\pmb{u}\|$. To address communication constraints we provide
deterministic and stochastic gradient compression schemes and show that with
these compression schemes our algorithm has worst-case optimal regret for the
case that all agents communicate in every round. Additionally, we exploit the
comparator-adaptive property of our algorithm to learn the best partition from
a set of candidate partitions, which allows different subsets of agents to
learn a different comparator.
Related papers
- Federated Contextual Cascading Bandits with Asynchronous Communication
and Heterogeneous Users [95.77678166036561]
We propose a UCB-type algorithm with delicate communication protocols.
We give sub-linear regret bounds on par with those achieved in the synchronous framework.
Empirical evaluation on synthetic and real-world datasets validates our algorithm's superior performance in terms of regrets and communication costs.
arXiv Detail & Related papers (2024-02-26T05:31:14Z) - Decentralized Multi-Task Online Convex Optimization Under Random Link
Failures [5.513958040574729]
We develop a robust decentralized saddle-point algorithm against random link failures with heterogeneous probabilities.
We extend our algorithm and analysis to the two-point bandit feedback scenario.
arXiv Detail & Related papers (2024-01-04T00:57:33Z) - Compressed Regression over Adaptive Networks [58.79251288443156]
We derive the performance achievable by a network of distributed agents that solve, adaptively and in the presence of communication constraints, a regression problem.
We devise an optimized allocation strategy where the parameters necessary for the optimization can be learned online by the agents.
arXiv Detail & Related papers (2023-04-07T13:41:08Z) - On Regret-optimal Cooperative Nonstochastic Multi-armed Bandits [7.23389716633927]
We show that a collaborative multi-agent emphfollow-the-regularized-leader (FTRL) algorithm has an individual regret upper bound that matches the lower bound up to a constant factor.
We also show that an FTRL algorithm with a suitable regularizer is regret optimal with respect to the scaling with the edge-delay parameter.
arXiv Detail & Related papers (2022-11-30T16:46:41Z) - Regret Minimization and Convergence to Equilibria in General-sum Markov
Games [57.568118148036376]
We present the first algorithm for learning in general-sum Markov games that provides sublinear regret guarantees when executed by all agents.
Our algorithm is decentralized, computationally efficient, and does not require any communication between agents.
arXiv Detail & Related papers (2022-07-28T16:27:59Z) - Contextual Model Aggregation for Fast and Robust Federated Learning in
Edge Computing [88.76112371510999]
Federated learning is a prime candidate for distributed machine learning at the network edge.
Existing algorithms face issues with slow convergence and/or robustness of performance.
We propose a contextual aggregation scheme that achieves the optimal context-dependent bound on loss reduction.
arXiv Detail & Related papers (2022-03-23T21:42:31Z) - Distributed Adaptive Learning Under Communication Constraints [54.22472738551687]
This work examines adaptive distributed learning strategies designed to operate under communication constraints.
We consider a network of agents that must solve an online optimization problem from continual observation of streaming data.
arXiv Detail & Related papers (2021-12-03T19:23:48Z) - An Efficient Algorithm for Cooperative Semi-Bandits [0.0]
We introduce Coop-FTPL, a cooperative version of the well-known Follow The Perturbed Leader algorithm.
We show that the expected regret of our algorithm after T time steps is of order QT log(k)(k$alpha$ 1 /Q + m), where Q is the total activation probability mass.
arXiv Detail & Related papers (2020-10-05T07:08:26Z) - Comparator-adaptive Convex Bandits [77.43350984086119]
We develop convex bandit algorithms with regret bounds that are small whenever the norm of the comparator is small.
We extend the ideas to convex bandits with Lipschitz or smooth loss functions.
arXiv Detail & Related papers (2020-07-16T16:33:35Z) - The Gossiping Insert-Eliminate Algorithm for Multi-Agent Bandits [20.259428328004738]
We consider a decentralized multi-agent Multi Armed Bandit (MAB) setup consisting of $N$ agents.
In our model, agents collaborate by exchanging messages through pairwise gossip style communications on an arbitrary connected graph.
arXiv Detail & Related papers (2020-01-15T17:49:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.