强化学习
计算机科学
最优控制
动态规划
人工神经网络
趋同(经济学)
数学优化
模型预测控制
电力系统
贝尔曼方程
功能(生物学)
集合(抽象数据类型)
功率(物理)
控制(管理)
人工智能
数学
算法
物理
经济
生物
进化生物学
量子力学
程序设计语言
经济增长
作者
Tong Wu,Anna Scaglione,Daniel Arnold
标识
DOI:10.1109/tpwrs.2023.3326121
摘要
Deep Reinforcement Learning (DRL) has emerged as a favored approach for resolving control challenges in power systems. Traditional DRL guides the agent through exploration of numerous policies, each embedded within a neural network (NN), aiming to maximize the associated reward function. However, this approach can lead to infeasible solutions that violate physical constraints such as power flow equations, voltage limits, and dynamic constraints. Ensuring these constraints are met is crucial in power systems, as they are a safety critical infrastructure. To address this issue, existing DRL algorithms remedy the problem by projecting the actions onto the feasible set, which can result in sub-optimal solutions. This paper introduces a pioneering primal-dual approach to learn optimal constrained DRL policies specifically for predictive control in real-time stochastic dynamic optimal power flow. The focus is on controlling power generations and battery outputs while ensuring compliance with critical constraints. We also prove the convergence of the critic and actor networks. Our case studies, based on IEEE standard systems, underscore the preeminence of the proposed approach in identifying near-optimal actions for various states while concurrently adhering to safety constraints.
科研通智能强力驱动
Strongly Powered by AbleSci AI