Multi-Agent Distributed Reinforcement Learning for Making Decentralized
Offloading Decisions
- URL: http://arxiv.org/abs/2204.02267v1
- Date: Tue, 5 Apr 2022 15:01:48 GMT
- Title: Multi-Agent Distributed Reinforcement Learning for Making Decentralized
Offloading Decisions
- Authors: Jing Tan and Ramin Khalili and Holger Karl and Artur Hecker
- Abstract summary: We formulate computation offloading as a decentralized decision-making problem with autonomous agents.
We design an interaction mechanism that incentivizes agents to align private and system goals by balancing between competition and cooperation.
For a dynamic environment, we propose a novel multi-agent online learning algorithm that learns with partial, delayed and noisy state information.
- Score: 7.326507804995567
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We formulate computation offloading as a decentralized decision-making
problem with autonomous agents. We design an interaction mechanism that
incentivizes agents to align private and system goals by balancing between
competition and cooperation. The mechanism provably has Nash equilibria with
optimal resource allocation in the static case. For a dynamic environment, we
propose a novel multi-agent online learning algorithm that learns with partial,
delayed and noisy state information, and a reward signal that reduces
information need to a great extent. Empirical results confirm that through
learning, agents significantly improve both system and individual performance,
e.g., 40% offloading failure rate reduction, 32% communication overhead
reduction, up to 38% computation resource savings in low contention, 18%
utilization increase with reduced load variation in high contention, and
improvement in fairness. Results also confirm the algorithm's good convergence
and generalization property in significantly different environments.
Related papers
- Boosting the Performance of Decentralized Federated Learning via Catalyst Acceleration [66.43954501171292]
We introduce Catalyst Acceleration and propose an acceleration Decentralized Federated Learning algorithm called DFedCata.
DFedCata consists of two main components: the Moreau envelope function, which addresses parameter inconsistencies, and Nesterov's extrapolation step, which accelerates the aggregation phase.
Empirically, we demonstrate the advantages of the proposed algorithm in both convergence speed and generalization performance on CIFAR10/100 with various non-iid data distributions.
arXiv Detail & Related papers (2024-10-09T06:17:16Z) - HERL: Tiered Federated Learning with Adaptive Homomorphic Encryption using Reinforcement Learning [12.628921853388862]
HERL is a Reinforcement Learning-based approach that uses Q-Learning to dynamically optimize encryption parameters.
Our proposed method involves first profiling and tiering clients according to the chosen clustering approach.
Our results show that HERL improves utility by 17%, reduces the convergence time by up to 24%, and increases efficiency by up to 30%, with minimal security loss.
arXiv Detail & Related papers (2024-09-11T21:26:23Z) - Multi-agent Assessment with QoS Enhancement for HD Map Updates in a Vehicular Network [8.853779271331508]
Reinforcement Learning (RL) algorithms have been used to address the challenging problems in the offloading process of vehicular ad hoc networks (VANET)
Deep Q-learning (DQN) and Actor-critic at the autonomous vehicle (AV) may lead to an increase in the computational load, causing a heavy burden on the computational devices and higher costs.
In this paper, we assess the scalability of an application utilizing a Q-learning single-agent solution in a distributed multi-agent environment.
arXiv Detail & Related papers (2024-07-31T09:17:09Z) - Asynchronous Federated Learning with Incentive Mechanism Based on
Contract Theory [5.502596101979607]
We propose a novel asynchronous FL framework that integrates an incentive mechanism based on contract theory.
Our framework exhibits a 1.35% accuracy improvement over the ideal Local SGD under attacks.
arXiv Detail & Related papers (2023-10-10T09:17:17Z) - Compressed Regression over Adaptive Networks [58.79251288443156]
We derive the performance achievable by a network of distributed agents that solve, adaptively and in the presence of communication constraints, a regression problem.
We devise an optimized allocation strategy where the parameters necessary for the optimization can be learned online by the agents.
arXiv Detail & Related papers (2023-04-07T13:41:08Z) - Multi-Agent Reinforcement Learning for Long-Term Network Resource
Allocation through Auction: a V2X Application [7.326507804995567]
We formulate offloading of computational tasks from a dynamic group of mobile agents (e.g., cars) as decentralized decision making among autonomous agents.
We design an interaction mechanism that incentivizes such agents to align private and system goals by balancing between competition and cooperation.
We propose a novel multi-agent online learning algorithm that learns with partial, delayed and noisy state information.
arXiv Detail & Related papers (2022-07-29T10:29:06Z) - Inducing Equilibria via Incentives: Simultaneous Design-and-Play Finds
Global Optima [114.31577038081026]
We propose an efficient method that tackles the designer's and agents' problems simultaneously in a single loop.
Although the designer does not solve the equilibrium problem repeatedly, it can anticipate the overall influence of the incentives on the agents.
We prove that the algorithm converges to the global optima at a sublinear rate for a broad class of games.
arXiv Detail & Related papers (2021-10-04T06:53:59Z) - Federated Learning with Unreliable Clients: Performance Analysis and
Mechanism Design [76.29738151117583]
Federated Learning (FL) has become a promising tool for training effective machine learning models among distributed clients.
However, low quality models could be uploaded to the aggregator server by unreliable clients, leading to a degradation or even a collapse of training.
We model these unreliable behaviors of clients and propose a defensive mechanism to mitigate such a security risk.
arXiv Detail & Related papers (2021-05-10T08:02:27Z) - Optimizing Mixed Autonomy Traffic Flow With Decentralized Autonomous
Vehicles and Multi-Agent RL [63.52264764099532]
We study the ability of autonomous vehicles to improve the throughput of a bottleneck using a fully decentralized control scheme in a mixed autonomy setting.
We apply multi-agent reinforcement algorithms to this problem and demonstrate that significant improvements in bottleneck throughput, from 20% at a 5% penetration rate to 33% at a 40% penetration rate, can be achieved.
arXiv Detail & Related papers (2020-10-30T22:06:05Z) - Value Variance Minimization for Learning Approximate Equilibrium in
Aggregation Systems [8.140037969280716]
We consider the problem of learning approximate equilibrium solutions (win-win) in aggregation systems.
In this paper, we consider the problem of learning approximate equilibrium solutions (win-win) in aggregation systems so that individuals have an incentive to remain in the aggregation system.
arXiv Detail & Related papers (2020-03-16T10:02:42Z) - Dynamic Federated Learning [57.14673504239551]
Federated learning has emerged as an umbrella term for centralized coordination strategies in multi-agent environments.
We consider a federated learning model where at every iteration, a random subset of available agents perform local updates based on their data.
Under a non-stationary random walk model on the true minimizer for the aggregate optimization problem, we establish that the performance of the architecture is determined by three factors, namely, the data variability at each agent, the model variability across all agents, and a tracking term that is inversely proportional to the learning rate of the algorithm.
arXiv Detail & Related papers (2020-02-20T15:00:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.