File Download
There are no files associated with this item.
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.1109/PESGM48719.2022.9917031
- Scopus: eid_2-s2.0-85141493880
Supplementary
-
Citations:
- Scopus: 0
- Appears in Collections:
Conference Paper: A Multi-Agent Reinforcement Learning based Frequency Control Method with Data-Enabled Predictive Control Guided Policy Search
Title | A Multi-Agent Reinforcement Learning based Frequency Control Method with Data-Enabled Predictive Control Guided Policy Search |
---|---|
Authors | |
Keywords | Data-enabled predictive control load frequency control multi-agent reinforcement learning multi-agent twin delayed deep deterministic policy gradient |
Issue Date | 17-Jul-2022 |
Abstract | In this paper, we propose a data-enabled predictive control (DeePC) guided multi-agent reinforcement learning (MARL) control algorithm, to solve the load frequency control (LFC) problem of power systems with multiple control areas. In the proposed algorithm, the offline training phase of MARL is guided by the DeePC algorithm. Moreover, a multi-agent twin delayed deep deterministic policy gradient (MATD3) algorithm is applied for LFC to reduce the overestimation bias of the Q-value in the existing multi-agent deep deterministic policy gradient (MADDPG) algorithm. Finally, the online control signals are generated directly by the trained policy neural networks. Simulation results on a two-area power system show that with the proposed DeePC-guided MATD3 algorithm: for one thing, the sample efficiency can be significantly improved, and thus the offline training phase can be stabilized and accelerated; for another thing, effective online cooperation of LFC between different control areas can be achieved. |
Persistent Identifier | http://hdl.handle.net/10722/333871 |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Zhao, Yunzheng | - |
dc.contributor.author | Liu, Tao | - |
dc.contributor.author | Hill, David J | - |
dc.date.accessioned | 2023-10-06T08:39:46Z | - |
dc.date.available | 2023-10-06T08:39:46Z | - |
dc.date.issued | 2022-07-17 | - |
dc.identifier.uri | http://hdl.handle.net/10722/333871 | - |
dc.description.abstract | <p>In this paper, we propose a data-enabled predictive control (DeePC) guided multi-agent reinforcement learning (MARL) control algorithm, to solve the load frequency control (LFC) problem of power systems with multiple control areas. In the proposed algorithm, the offline training phase of MARL is guided by the DeePC algorithm. Moreover, a multi-agent twin delayed deep deterministic policy gradient (MATD3) algorithm is applied for LFC to reduce the overestimation bias of the Q-value in the existing multi-agent deep deterministic policy gradient (MADDPG) algorithm. Finally, the online control signals are generated directly by the trained policy neural networks. Simulation results on a two-area power system show that with the proposed DeePC-guided MATD3 algorithm: for one thing, the sample efficiency can be significantly improved, and thus the offline training phase can be stabilized and accelerated; for another thing, effective online cooperation of LFC between different control areas can be achieved.<br></p> | - |
dc.language | eng | - |
dc.relation.ispartof | IEEE Power & Energy Society General Meeting (17/07/2022-21/07/2022, Denver) | - |
dc.subject | Data-enabled predictive control | - |
dc.subject | load frequency control | - |
dc.subject | multi-agent reinforcement learning | - |
dc.subject | multi-agent twin delayed deep deterministic policy gradient | - |
dc.title | A Multi-Agent Reinforcement Learning based Frequency Control Method with Data-Enabled Predictive Control Guided Policy Search | - |
dc.type | Conference_Paper | - |
dc.identifier.doi | 10.1109/PESGM48719.2022.9917031 | - |
dc.identifier.scopus | eid_2-s2.0-85141493880 | - |
dc.identifier.volume | 2022-July | - |