Demand-Side Scheduling Based on Multi-Agent Deep Actor-Critic Learning
for Smart Grids
- URL: http://arxiv.org/abs/2005.01979v2
- Date: Tue, 23 Aug 2022 10:12:58 GMT
- Title: Demand-Side Scheduling Based on Multi-Agent Deep Actor-Critic Learning
for Smart Grids
- Authors: Joash Lee, Wenbo Wang, Dusit Niyato
- Abstract summary: We consider the problem of demand-side energy management, where each household is equipped with a smart meter that is able to schedule home appliances online.
The goal is to minimize the overall cost under a real-time pricing scheme.
We propose the formulation of a smart grid environment as a Markov game.
- Score: 56.35173057183362
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We consider the problem of demand-side energy management, where each
household is equipped with a smart meter that is able to schedule home
appliances online. The goal is to minimize the overall cost under a real-time
pricing scheme. While previous works have introduced centralized approaches in
which the scheduling algorithm has full observability, we propose the
formulation of a smart grid environment as a Markov game. Each household is a
decentralized agent with partial observability, which allows scalability and
privacy-preservation in a realistic setting. The grid operator produces a price
signal that varies with the energy demand. We propose an extension to a
multi-agent, deep actor-critic algorithm to address partial observability and
the perceived non-stationarity of the environment from the agent's viewpoint.
This algorithm learns a centralized critic that coordinates training of
decentralized agents. Our approach thus uses centralized learning but
decentralized execution. Simulation results show that our online deep
reinforcement learning method can reduce both the peak-to-average ratio of
total energy consumed and the cost of electricity for all households based
purely on instantaneous observations and a price signal.
Related papers
- Generalizability of Graph Neural Networks for Decentralized Unlabeled Motion Planning [72.86540018081531]
Unlabeled motion planning involves assigning a set of robots to target locations while ensuring collision avoidance.
This problem forms an essential building block for multi-robot systems in applications such as exploration, surveillance, and transportation.
We address this problem in a decentralized setting where each robot knows only the positions of its $k$-nearest robots and $k$-nearest targets.
arXiv Detail & Related papers (2024-09-29T23:57:25Z) - Decentralized Coordination of Distributed Energy Resources through Local Energy Markets and Deep Reinforcement Learning [1.8434042562191815]
Transactive energy, implemented through local energy markets, has recently garnered attention as a promising solution to the grid challenges.
This study addresses the gap by training a set of deep reinforcement learning agents to automate end-user participation in ALEX.
The study unveils a clear correlation between bill reduction and reduced net load variability in this setup.
arXiv Detail & Related papers (2024-04-19T19:03:33Z) - Centralized vs. Decentralized Multi-Agent Reinforcement Learning for Enhanced Control of Electric Vehicle Charging Networks [1.9188272016043582]
We introduce a novel approach for distributed and cooperative charging strategy using a Multi-Agent Reinforcement Learning (MARL) framework.
Our method is built upon the Deep Deterministic Policy Gradient (DDPG) algorithm for a group of EVs in a residential community.
Our results indicate that, despite higher policy variances and training complexity, the CTDE-DDPG framework significantly improves charging efficiency by reducing total variation by approximately %36 and charging cost by around %9.1 on average.
arXiv Detail & Related papers (2024-04-18T21:50:03Z) - TransPath: Learning Heuristics For Grid-Based Pathfinding via
Transformers [64.88759709443819]
We suggest learning the instance-dependent proxies that are supposed to notably increase the efficiency of the search.
The first proxy we suggest to learn is the correction factor, i.e. the ratio between the instance independent cost-to-go estimate and the perfect one.
The second proxy is the path probability, which indicates how likely the grid cell is lying on the shortest path.
arXiv Detail & Related papers (2022-12-22T14:26:11Z) - Adaptive Stochastic ADMM for Decentralized Reinforcement Learning in
Edge Industrial IoT [106.83952081124195]
Reinforcement learning (RL) has been widely investigated and shown to be a promising solution for decision-making and optimal control processes.
We propose an adaptive ADMM (asI-ADMM) algorithm and apply it to decentralized RL with edge-computing-empowered IIoT networks.
Experiment results show that our proposed algorithms outperform the state of the art in terms of communication costs and scalability, and can well adapt to complex IoT environments.
arXiv Detail & Related papers (2021-06-30T16:49:07Z) - A Learning-based Optimal Market Bidding Strategy for Price-Maker Energy
Storage [3.0839245814393728]
We implement an online Supervised Actor-Critic (SAC) algorithm supervised with a model-based controller -- Model Predictive Control (MPC)
The energy storage agent is trained with this algorithm to optimally bid while learning and adjusting to its impact on the market clearing prices.
Our contribution, thus, is an online and safe SAC algorithm that outperforms the current model-based state-of-the-art.
arXiv Detail & Related papers (2021-06-04T10:22:58Z) - Multi-agent Reinforcement Learning for Resource Allocation in IoT
networks with Edge Computing [16.129649374251088]
It's challenging for end users to offload computation due to their massive requirements on spectrum and resources.
In this paper, we investigate offloading mechanism with resource allocation in IoT edge computing networks by formulating it as a game.
arXiv Detail & Related papers (2020-04-05T20:59:20Z) - Risk-Aware Energy Scheduling for Edge Computing with Microgrid: A
Multi-Agent Deep Reinforcement Learning Approach [82.6692222294594]
We study a risk-aware energy scheduling problem for a microgrid-powered MEC network.
We derive the solution by applying a multi-agent deep reinforcement learning (MADRL)-based advantage actor-critic (A3C) algorithm with shared neural networks.
arXiv Detail & Related papers (2020-02-21T02:14:38Z) - Multi-Agent Meta-Reinforcement Learning for Self-Powered and Sustainable
Edge Computing Systems [87.4519172058185]
An effective energy dispatch mechanism for self-powered wireless networks with edge computing capabilities is studied.
A novel multi-agent meta-reinforcement learning (MAMRL) framework is proposed to solve the formulated problem.
Experimental results show that the proposed MAMRL model can reduce up to 11% non-renewable energy usage and by 22.4% the energy cost.
arXiv Detail & Related papers (2020-02-20T04:58:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.