软件部署
强化学习
计算机科学
弹性(材料科学)
分布式计算
节点(物理)
人口
事件(粒子物理)
航程(航空)
电力系统
功率(物理)
人工智能
工程类
结构工程
热力学
航空航天工程
社会学
量子力学
人口学
操作系统
物理
作者
Mukesh Gautam,Narayan Bhusal,Mohammed Benidris
标识
DOI:10.1109/mias.2023.3325046
摘要
The deployment of movable energy resources (MERs) can be an effective strategy to restore critical loads to enhance power system resilience when no other energy sources are available after the occurrence of an extreme event. Since the optimal locations of MERs following an extreme event are dependent on system operating states (e.g., the loads at each node, on/off status of system branches, and so on), existing analytical and population-based approaches must repeat the entire analysis and calculation when the system operating states change. On the contrary, if deep reinforcement learning (DRL)-based algorithms are sufficiently trained with a wide range of scenarios, they can quickly find optimal or near-optimal locations irrespective of changes in system states. A deep Q-learning-based approach is proposed for optimal MER deployment to enhance power system resilience. MERs can be also utilized to complement other types of resources, if available. The proposed approach operates in two stages after the occurrence of extreme events. In the first stage, the distribution network is represented as a graph, and the network is then reconfigured using tie switches by using Kruskal’s spanning forest search algorithm (KSFSA). To maximize critical load recovery, the optimal or near-optimal locations of MERs are chosen in the second stage. Further, case studies on a 33-node distribution system and a modified IEEE 123-node system demonstrate the effectiveness of the proposed approach for postdisaster routing of MERs.
科研通智能强力驱动
Strongly Powered by AbleSci AI