Flooding with Absorption: An Efficient Protocol for Heterogeneous
Bandits over Complex Networks
- URL: http://arxiv.org/abs/2303.05445v4
- Date: Sun, 25 Feb 2024 12:05:43 GMT
- Title: Flooding with Absorption: An Efficient Protocol for Heterogeneous
Bandits over Complex Networks
- Authors: Junghyun Lee, Laura Schmid, Se-Young Yun
- Abstract summary: We consider a multi-agent setting where each agent solves their own bandit instance endowed with a different set of arms.
Their goal is to minimize their group regret while collaborating via some communication protocol over a given network.
We propose a new protocol called Flooding with Absorption (FwA) to mitigate the issue of high communication costs incurred by flooding in complex networks.
- Score: 30.94416632071414
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Multi-armed bandits are extensively used to model sequential decision-making,
making them ubiquitous in many real-life applications such as online
recommender systems and wireless networking. We consider a multi-agent setting
where each agent solves their own bandit instance endowed with a different set
of arms. Their goal is to minimize their group regret while collaborating via
some communication protocol over a given network. Previous literature on this
problem only considered arm heterogeneity and networked agents separately. In
this work, we introduce a setting that encompasses both features. For this
novel setting, we first provide a rigorous regret analysis for a standard
flooding protocol combined with the classic UCB policy. Then, to mitigate the
issue of high communication costs incurred by flooding in complex networks, we
propose a new protocol called Flooding with Absorption (FwA). We provide a
theoretical analysis of the resulting regret bound and discuss the advantages
of using FwA over flooding. Lastly, we experimentally verify on various
scenarios, including dynamic networks, that FwA leads to significantly lower
communication costs despite minimal regret performance loss compared to other
network protocols.
Related papers
- Leveraging Low-Rank and Sparse Recurrent Connectivity for Robust
Closed-Loop Control [63.310780486820796]
We show how a parameterization of recurrent connectivity influences robustness in closed-loop settings.
We find that closed-form continuous-time neural networks (CfCs) with fewer parameters can outperform their full-rank, fully-connected counterparts.
arXiv Detail & Related papers (2023-10-05T21:44:18Z) - Communication-Efficient Zeroth-Order Distributed Online Optimization:
Algorithm, Theory, and Applications [9.045332526072828]
This paper focuses on a multi-agent zeroth-order online optimization problem in a federated learning setting for target tracking.
The proposed solution is further analyzed in terms of errors and errors in two relevant applications.
arXiv Detail & Related papers (2023-06-09T03:51:45Z) - Compressed Regression over Adaptive Networks [58.79251288443156]
We derive the performance achievable by a network of distributed agents that solve, adaptively and in the presence of communication constraints, a regression problem.
We devise an optimized allocation strategy where the parameters necessary for the optimization can be learned online by the agents.
arXiv Detail & Related papers (2023-04-07T13:41:08Z) - The Double-Edged Sword of Implicit Bias: Generalization vs. Robustness
in ReLU Networks [64.12052498909105]
We study the implications of the implicit bias of gradient flow on generalization and adversarial robustness in ReLU networks.
In two-layer ReLU networks gradient flow is biased towards solutions that generalize well, but are highly vulnerable to adversarial examples.
arXiv Detail & Related papers (2023-03-02T18:14:35Z) - On Regret-optimal Cooperative Nonstochastic Multi-armed Bandits [7.23389716633927]
We show that a collaborative multi-agent emphfollow-the-regularized-leader (FTRL) algorithm has an individual regret upper bound that matches the lower bound up to a constant factor.
We also show that an FTRL algorithm with a suitable regularizer is regret optimal with respect to the scaling with the edge-delay parameter.
arXiv Detail & Related papers (2022-11-30T16:46:41Z) - Manifold Regularized Dynamic Network Pruning [102.24146031250034]
This paper proposes a new paradigm that dynamically removes redundant filters by embedding the manifold information of all instances into the space of pruned networks.
The effectiveness of the proposed method is verified on several benchmarks, which shows better performance in terms of both accuracy and computational cost.
arXiv Detail & Related papers (2021-03-10T03:59:03Z) - ADOM: Accelerated Decentralized Optimization Method for Time-Varying
Networks [124.33353902111939]
We propose ADOM - an accelerated method for smooth and strongly convex decentralized optimization over time-varying networks.
Up to a constant factor, which depends on the network structure only, its communication is the same as that of accelerated Nesterov method.
arXiv Detail & Related papers (2021-02-18T09:37:20Z) - Decentralized Multi-Agent Linear Bandits with Safety Constraints [31.67685495996986]
We study decentralized linear bandits, where a network of $N$ agents acts cooperatively to solve a linear bandit-optimization problem.
We propose DLUCB: a fully decentralized algorithm that minimizes the cumulative regret over the entire network.
We show that our ideas extend naturally to the emerging, albeit more challenging, setting of safe bandits.
arXiv Detail & Related papers (2020-12-01T07:33:00Z) - ESPN: Extremely Sparse Pruned Networks [50.436905934791035]
We show that a simple iterative mask discovery method can achieve state-of-the-art compression of very deep networks.
Our algorithm represents a hybrid approach between single shot network pruning methods and Lottery-Ticket type approaches.
arXiv Detail & Related papers (2020-06-28T23:09:27Z) - Coagent Networks Revisited [10.45819881530349]
Coagent networks formalize the concept of arbitrary networks of agents that collaborate to take actions in a reinforcement learning environment.
We first provide a unifying perspective on the many diverse examples that fall under coagent networks.
We do so by formalizing the rules of execution in a coagent network, enabled by the novel and intuitive idea of execution paths.
arXiv Detail & Related papers (2020-01-28T17:31:23Z) - Information Compensation for Deep Conditional Generative Networks [38.054911004694624]
We propose a novel structure for unsupervised conditional GANs powered by a novel Information Compensation Connection (IC-Connection)
The proposed IC-Connection enables GANs to compensate for information loss incurred during deconvolution operations.
Our empirical results suggest that our method achieves better disentanglement compared to the state-of-the-art GANs in a conditional generation setting.
arXiv Detail & Related papers (2020-01-23T14:39:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.