强化学习
马尔可夫决策过程
稳健性(进化)
自动频率控制
可扩展性
计算机科学
增强学习
电力系统
网格
马尔可夫过程
控制工程
人工智能
工程类
功率(物理)
电信
几何学
量子力学
化学
物理
数学
统计
基因
生物化学
数据库
标识
DOI:10.1109/tpwrs.2021.3130413
摘要
Emergency frequency control is one of the most critical approaches to maintain power system stability after major disturbances. With the increasing number of grid-connected renewable energy sources, existing model-based methods of frequency control are facing up with challenges of computational speed and scalability for large-scale systems. In this paper, the emergency frequency control problem is formulated as a Markov Decision Process and solved through a novel distributional deep reinforcement learning (DRL) method, namely the distributional soft actor critic (DSAC) method. Compared with other reinforcement learning methods that only estimate the mean value, the proposed DSAC model estimates the distribution of value function over returns. This advancement can lead to more insights and knowledge for the agent, with the benefit of a much faster and more stable learning process, and the improved frequency control performance. The simulation results on IEEE 39-bus and IEEE 118-bus systems demonstrate the effectiveness and robustness of proposed models, as well as the advantage compared to other state-of-the-art DRL algorithms.
科研通智能强力驱动
Strongly Powered by AbleSci AI