Multi-agent Bayesian Deep Reinforcement Learning for Microgrid Energy
Management under Communication Failures
- URL: http://arxiv.org/abs/2111.11868v1
- Date: Mon, 22 Nov 2021 03:08:10 GMT
- Title: Multi-agent Bayesian Deep Reinforcement Learning for Microgrid Energy
Management under Communication Failures
- Authors: Hao Zhou, Atakan Aral, Ivona Brandic, Melike Erol-Kantarci
- Abstract summary: We propose a multi-agent Bayesian deep reinforcement learning (BA-DRL) method for MG energy management under communication failures.
BA-DRL has 4.1% and 10.3% higher reward than Nash Deep Q-learning (Nash-DQN) and alternating direction method of multipliers (ADMM) respectively under 1% communication failure probability.
- Score: 10.099371194251052
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Microgrids (MGs) are important players for the future transactive energy
systems where a number of intelligent Internet of Things (IoT) devices interact
for energy management in the smart grid. Although there have been many works on
MG energy management, most studies assume a perfect communication environment,
where communication failures are not considered. In this paper, we consider the
MG as a multi-agent environment with IoT devices in which AI agents exchange
information with their peers for collaboration. However, the collaboration
information may be lost due to communication failures or packet loss. Such
events may affect the operation of the whole MG. To this end, we propose a
multi-agent Bayesian deep reinforcement learning (BA-DRL) method for MG energy
management under communication failures. We first define a multi-agent
partially observable Markov decision process (MA-POMDP) to describe agents
under communication failures, in which each agent can update its beliefs on the
actions of its peers. Then, we apply a double deep Q-learning (DDQN)
architecture for Q-value estimation in BA-DRL, and propose a belief-based
correlated equilibrium for the joint-action selection of multi-agent BA-DRL.
Finally, the simulation results show that BA-DRL is robust to both power supply
uncertainty and communication failure uncertainty. BA-DRL has 4.1% and 10.3%
higher reward than Nash Deep Q-learning (Nash-DQN) and alternating direction
method of multipliers (ADMM) respectively under 1% communication failure
probability.
Related papers
- Large Language Model Enhanced Multi-Agent Systems for 6G Communications [94.45712802626794]
We propose a multi-agent system with customized communication knowledge and tools for solving communication related tasks using natural language.
We validate the effectiveness of the proposed multi-agent system by designing a semantic communication system.
arXiv Detail & Related papers (2023-12-13T02:35:57Z) - Building Cooperative Embodied Agents Modularly with Large Language
Models [104.57849816689559]
We address challenging multi-agent cooperation problems with decentralized control, raw sensory observations, costly communication, and multi-objective tasks instantiated in various embodied environments.
We harness the commonsense knowledge, reasoning ability, language comprehension, and text generation prowess of LLMs and seamlessly incorporate them into a cognitive-inspired modular framework.
Our experiments on C-WAH and TDW-MAT demonstrate that CoELA driven by GPT-4 can surpass strong planning-based methods and exhibit emergent effective communication.
arXiv Detail & Related papers (2023-07-05T17:59:27Z) - Federated Multi-Agent Deep Reinforcement Learning Approach via
Physics-Informed Reward for Multi-Microgrid Energy Management [34.18923657108073]
This paper proposes a federated multi-agent deep reinforcement learning (F-MADRL) algorithm via the physics-informed reward.
In this algorithm, the federated learning mechanism is introduced to train the F-MADRL algorithm thus ensures the privacy and the security of data.
Experiments are conducted on Oak Ridge national laboratory distributed energy control communication lab microgrid (ORNL-MG) test system.
arXiv Detail & Related papers (2022-12-29T08:35:11Z) - Provably Efficient Multi-Agent Reinforcement Learning with Fully
Decentralized Communication [3.5450828190071655]
Distributed exploration reduces sampling complexity in reinforcement learning.
We show that group performance can be significantly improved when each agent uses a decentralized message-passing protocol.
We show that incorporating more agents and more information sharing into the group learning scheme speeds up convergence to the optimal policy.
arXiv Detail & Related papers (2021-10-14T14:27:27Z) - Correlated Deep Q-learning based Microgrid Energy Management [12.013067383415747]
This paper proposes a correlated deep Q-learning (CDQN) based technique for the MG energy management.
Each electrical entity is modeled as an agent which has a neural network to predict its own Q-values, after which the correlated Q-equilibrium is used to coordinate agents.
The simulation result shows 40.9% and 9.62% higher profit for ESS agent and photovoltaic (PV) agent, respectively.
arXiv Detail & Related papers (2021-03-06T16:43:18Z) - Information Freshness-Aware Task Offloading in Air-Ground Integrated
Edge Computing Systems [49.80033982995667]
This paper studies the problem of information freshness-aware task offloading in an air-ground integrated multi-access edge computing system.
A third-party real-time application service provider provides computing services to the subscribed mobile users (MUs) with the limited communication and computation resources from the InP.
We derive a novel deep reinforcement learning (RL) scheme that adopts two separate double deep Q-networks for each MU to approximate the Q-factor and the post-decision Q-factor.
arXiv Detail & Related papers (2020-07-15T21:32:43Z) - Learning Individually Inferred Communication for Multi-Agent Cooperation [37.56115000150748]
We propose Individually Inferred Communication (I2C) to enable agents to learn a prior for agent-agent communication.
The prior knowledge is learned via causal inference and realized by a feed-forward neural network.
I2C can not only reduce communication overhead but also improve the performance in a variety of multi-agent cooperative scenarios.
arXiv Detail & Related papers (2020-06-11T14:07:57Z) - Distributed Reinforcement Learning for Cooperative Multi-Robot Object
Manipulation [53.262360083572005]
We consider solving a cooperative multi-robot object manipulation task using reinforcement learning (RL)
We propose two distributed multi-agent RL approaches: distributed approximate RL (DA-RL) and game-theoretic RL (GT-RL)
Although we focus on a small system of two agents in this paper, both DA-RL and GT-RL apply to general multi-agent systems, and are expected to scale well to large systems.
arXiv Detail & Related papers (2020-03-21T00:43:54Z) - Monotonic Value Function Factorisation for Deep Multi-Agent
Reinforcement Learning [55.20040781688844]
QMIX is a novel value-based method that can train decentralised policies in a centralised end-to-end fashion.
We propose the StarCraft Multi-Agent Challenge (SMAC) as a new benchmark for deep multi-agent reinforcement learning.
arXiv Detail & Related papers (2020-03-19T16:51:51Z) - Multi-Agent Meta-Reinforcement Learning for Self-Powered and Sustainable
Edge Computing Systems [87.4519172058185]
An effective energy dispatch mechanism for self-powered wireless networks with edge computing capabilities is studied.
A novel multi-agent meta-reinforcement learning (MAMRL) framework is proposed to solve the formulated problem.
Experimental results show that the proposed MAMRL model can reduce up to 11% non-renewable energy usage and by 22.4% the energy cost.
arXiv Detail & Related papers (2020-02-20T04:58:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.