Online Multi-agent Reinforcement Learning for Decentralized
Inverter-based Volt-VAR Control
- URL: http://arxiv.org/abs/2006.12841v2
- Date: Wed, 7 Apr 2021 15:31:23 GMT
- Title: Online Multi-agent Reinforcement Learning for Decentralized
Inverter-based Volt-VAR Control
- Authors: Haotian Liu, Wenchuan Wu
- Abstract summary: The distributed Volt/Var control (VVC) methods have been widely studied for active distribution networks(ADNs)
We propose an online multi-agent reinforcement learning and decentralized control framework (OLDC) for VVC.
- Score: 3.260913246106564
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The distributed Volt/Var control (VVC) methods have been widely studied for
active distribution networks(ADNs), which is based on perfect model and
real-time P2P communication. However, the model is always incomplete with
significant parameter errors and such P2P communication system is hard to
maintain. In this paper, we propose an online multi-agent reinforcement
learning and decentralized control framework (OLDC) for VVC. In this framework,
the VVC problem is formulated as a constrained Markov game and we propose a
novel multi-agent constrained soft actor-critic (MACSAC) reinforcement learning
algorithm. MACSAC is used to train the control agents online, so the accurate
ADN model is no longer needed. Then, the trained agents can realize
decentralized optimal control using local measurements without real-time P2P
communication. The OLDC with MACSAC has shown extraordinary flexibility,
efficiency and robustness to various computing and communication conditions.
Numerical simulations on IEEE test cases not only demonstrate that the proposed
MACSAC outperforms the state-of-art learning algorithms, but also support the
superiority of our OLDC framework in the online application.
Related papers
- Augmented Lagrangian-Based Safe Reinforcement Learning Approach for Distribution System Volt/VAR Control [1.1059341532498634]
This paper formulates the Volt- VAR control problem as a constrained Markov decision process (CMDP)
A novel safe off-policy reinforcement learning (RL) approach is proposed in this paper to solve the CMDP.
A two-stage strategy is adopted for offline training and online execution, so the accurate distribution system model is no longer needed.
arXiv Detail & Related papers (2024-10-19T19:45:09Z) - Parameter-Adaptive Approximate MPC: Tuning Neural-Network Controllers without Retraining [50.00291020618743]
This work introduces a novel, parameter-adaptive AMPC architecture capable of online tuning without recomputing large datasets and retraining.
We showcase the effectiveness of parameter-adaptive AMPC by controlling the swing-ups of two different real cartpole systems with a severely resource-constrained microcontroller (MCU)
Taken together, these contributions represent a marked step toward the practical application of AMPC in real-world systems.
arXiv Detail & Related papers (2024-04-08T20:02:19Z) - A Unified Framework for Alternating Offline Model Training and Policy
Learning [62.19209005400561]
In offline model-based reinforcement learning, we learn a dynamic model from historically collected data, and utilize the learned model and fixed datasets for policy learning.
We develop an iterative offline MBRL framework, where we maximize a lower bound of the true expected return.
With the proposed unified model-policy learning framework, we achieve competitive performance on a wide range of continuous-control offline reinforcement learning datasets.
arXiv Detail & Related papers (2022-10-12T04:58:51Z) - MACC: Cross-Layer Multi-Agent Congestion Control with Deep Reinforcement
Learning [14.29757990259669]
Congestion Control (CC) is core networking task to efficiently utilize network capacity.
In this paper, we explore the performance of multi-agent reinforcement learning-based cross-layer congestion control algorithms.
arXiv Detail & Related papers (2022-06-04T12:02:35Z) - Optimization for Master-UAV-powered Auxiliary-Aerial-IRS-assisted IoT
Networks: An Option-based Multi-agent Hierarchical Deep Reinforcement
Learning Approach [56.84948632954274]
This paper investigates a master unmanned aerial vehicle (MUAV)-powered Internet of Things (IoT) network.
We propose using a rechargeable auxiliary UAV (AUAV) equipped with an intelligent reflecting surface (IRS) to enhance the communication signals from the MUAV.
Under the proposed model, we investigate the optimal collaboration strategy of these energy-limited UAVs to maximize the accumulated throughput of the IoT network.
arXiv Detail & Related papers (2021-12-20T15:45:28Z) - Adaptive Stochastic ADMM for Decentralized Reinforcement Learning in
Edge Industrial IoT [106.83952081124195]
Reinforcement learning (RL) has been widely investigated and shown to be a promising solution for decision-making and optimal control processes.
We propose an adaptive ADMM (asI-ADMM) algorithm and apply it to decentralized RL with edge-computing-empowered IIoT networks.
Experiment results show that our proposed algorithms outperform the state of the art in terms of communication costs and scalability, and can well adapt to complex IoT environments.
arXiv Detail & Related papers (2021-06-30T16:49:07Z) - Lyapunov-Based Reinforcement Learning for Decentralized Multi-Agent
Control [3.3788926259119645]
In decentralized multi-agent control, systems are complex with unknown or highly uncertain dynamics.
Deep reinforcement learning (DRL) is promising to learn the controller/policy from data without the knowing system dynamics.
Existing multi-agent reinforcement learning (MARL) algorithms cannot ensure the closed-loop stability of a multi-agent system.
We propose a new MARL algorithm for decentralized multi-agent control with a stability guarantee.
arXiv Detail & Related papers (2020-09-20T06:11:42Z) - Consensus Multi-Agent Reinforcement Learning for Volt-VAR Control in
Power Distribution Networks [8.472603460083375]
We propose consensus multi-agent deep reinforcement learning algorithm to solve the VVC problem.
The proposed algorithm allows individual agents to learn a group control policy using local rewards.
Numerical studies on IEEE distribution test feeders show that our proposed algorithm matches the performance of single-agent reinforcement learning benchmark.
arXiv Detail & Related papers (2020-07-06T18:21:47Z) - Two-stage Deep Reinforcement Learning for Inverter-based Volt-VAR
Control in Active Distribution Networks [3.260913246106564]
We propose a novel two-stage deep reinforcement learning (DRL) method to improve the voltage profile by regulating inverter-based energy resources.
In the offline stage, a highly efficient adversarial reinforcement learning algorithm is developed to train an offline agent robust to the model mismatch.
In the sequential online stage, we transfer the offline agent safely as the online agent to perform continuous learning and controlling online with significantly improved safety and efficiency.
arXiv Detail & Related papers (2020-05-20T08:02:13Z) - F2A2: Flexible Fully-decentralized Approximate Actor-critic for
Cooperative Multi-agent Reinforcement Learning [110.35516334788687]
Decentralized multi-agent reinforcement learning algorithms are sometimes unpractical in complicated applications.
We propose a flexible fully decentralized actor-critic MARL framework, which can handle large-scale general cooperative multi-agent setting.
Our framework can achieve scalability and stability for large-scale environment and reduce information transmission.
arXiv Detail & Related papers (2020-04-17T14:56:29Z) - Information Theoretic Model Predictive Q-Learning [64.74041985237105]
We present a novel theoretical connection between information theoretic MPC and entropy regularized RL.
We develop a Q-learning algorithm that can leverage biased models.
arXiv Detail & Related papers (2019-12-31T00:29:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.