Distributed Adaptive Learning Under Communication Constraints
- URL: http://arxiv.org/abs/2112.02129v1
- Date: Fri, 3 Dec 2021 19:23:48 GMT
- Title: Distributed Adaptive Learning Under Communication Constraints
- Authors: Marco Carpentiero, Vincenzo Matta, Ali H. Sayed
- Abstract summary: This work examines adaptive distributed learning strategies designed to operate under communication constraints.
We consider a network of agents that must solve an online optimization problem from continual observation of streaming data.
- Score: 54.22472738551687
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This work examines adaptive distributed learning strategies designed to
operate under communication constraints. We consider a network of agents that
must solve an online optimization problem from continual observation of
streaming data. The agents implement a distributed cooperative strategy where
each agent is allowed to perform local exchange of information with its
neighbors. In order to cope with communication constraints, the exchanged
information must be unavoidably compressed. We propose a diffusion strategy
nicknamed as ACTC (Adapt-Compress-Then-Combine), which relies on the following
steps: i) an adaptation step where each agent performs an individual
stochastic-gradient update with constant step-size; ii) a compression step that
leverages a recently introduced class of stochastic compression operators; and
iii) a combination step where each agent combines the compressed updates
received from its neighbors. The distinguishing elements of this work are as
follows. First, we focus on adaptive strategies, where constant (as opposed to
diminishing) step-sizes are critical to respond in real time to nonstationary
variations. Second, we consider the general class of directed graphs and
left-stochastic combination policies, which allow us to enhance the interplay
between topology and learning. Third, in contrast with related works that
assume strong convexity for all individual agents' cost functions, we require
strong convexity only at a network level, a condition satisfied even if a
single agent has a strongly-convex cost and the remaining agents have
non-convex costs. Fourth, we focus on a diffusion (as opposed to consensus)
strategy. Under the demanding setting of compressed information, we establish
that the ACTC iterates fluctuate around the desired optimizer, achieving
remarkable savings in terms of bits exchanged between neighboring agents.
Related papers
- Communication Learning in Multi-Agent Systems from Graph Modeling Perspective [62.13508281188895]
We introduce a novel approach wherein we conceptualize the communication architecture among agents as a learnable graph.
We introduce a temporal gating mechanism for each agent, enabling dynamic decisions on whether to receive shared information at a given time.
arXiv Detail & Related papers (2024-11-01T05:56:51Z) - Learning Multi-Agent Communication from Graph Modeling Perspective [62.13508281188895]
We introduce a novel approach wherein we conceptualize the communication architecture among agents as a learnable graph.
Our proposed approach, CommFormer, efficiently optimize the communication graph and concurrently refines architectural parameters through gradient descent in an end-to-end manner.
arXiv Detail & Related papers (2024-05-14T12:40:25Z) - Decentralized Learning Strategies for Estimation Error Minimization with Graph Neural Networks [94.2860766709971]
We address the challenge of sampling and remote estimation for autoregressive Markovian processes in a wireless network with statistically-identical agents.
Our goal is to minimize time-average estimation error and/or age of information with decentralized scalable sampling and transmission policies.
arXiv Detail & Related papers (2024-04-04T06:24:11Z) - Communication-Efficient Zeroth-Order Distributed Online Optimization:
Algorithm, Theory, and Applications [9.045332526072828]
This paper focuses on a multi-agent zeroth-order online optimization problem in a federated learning setting for target tracking.
The proposed solution is further analyzed in terms of errors and errors in two relevant applications.
arXiv Detail & Related papers (2023-06-09T03:51:45Z) - Compressed Regression over Adaptive Networks [58.79251288443156]
We derive the performance achievable by a network of distributed agents that solve, adaptively and in the presence of communication constraints, a regression problem.
We devise an optimized allocation strategy where the parameters necessary for the optimization can be learned online by the agents.
arXiv Detail & Related papers (2023-04-07T13:41:08Z) - Depthwise Convolution for Multi-Agent Communication with Enhanced
Mean-Field Approximation [9.854975702211165]
We propose a new method based on local communication learning to tackle the multi-agent RL (MARL) challenge.
First, we design a new communication protocol that exploits the ability of depthwise convolution to efficiently extract local relations.
Second, we introduce the mean-field approximation into our method to reduce the scale of agent interactions.
arXiv Detail & Related papers (2022-03-06T07:42:43Z) - Dimension-Free Rates for Natural Policy Gradient in Multi-Agent
Reinforcement Learning [22.310861786709538]
We propose a scalable algorithm for cooperative multi-agent reinforcement learning.
We show that our algorithm converges to the globally optimal policy with a dimension-free statistical and computational complexity.
arXiv Detail & Related papers (2021-09-23T23:38:15Z) - ROMAX: Certifiably Robust Deep Multiagent Reinforcement Learning via
Convex Relaxation [32.091346776897744]
Cyber-physical attacks can challenge the robustness of multiagent reinforcement learning.
We propose a minimax MARL approach to infer the worst-case policy update of other agents.
arXiv Detail & Related papers (2021-09-14T16:18:35Z) - Dynamic Federated Learning [57.14673504239551]
Federated learning has emerged as an umbrella term for centralized coordination strategies in multi-agent environments.
We consider a federated learning model where at every iteration, a random subset of available agents perform local updates based on their data.
Under a non-stationary random walk model on the true minimizer for the aggregate optimization problem, we establish that the performance of the architecture is determined by three factors, namely, the data variability at each agent, the model variability across all agents, and a tracking term that is inversely proportional to the learning rate of the algorithm.
arXiv Detail & Related papers (2020-02-20T15:00:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.