增广拉格朗日法
马尔可夫决策过程
强化学习
数学优化
调度(生产过程)
计算机科学
电动汽车
需求响应
灵活性(工程)
拉格朗日
拉格朗日松弛
算法
功率(物理)
马尔可夫过程
工程类
电
数学
人工智能
统计
物理
量子力学
电气工程
数学物理
作者
Xiaoying Shi,Yinliang Xu,Guibin Chen,Ye Guo
标识
DOI:10.1109/tsg.2023.3289211
摘要
This paper proposes an augmented Lagrangian-based safe off-policy deep reinforcement learning (DRL) algorithm for the carbon-oriented optimal scheduling of electric vehicle (EV) aggregators in a distribution network. First, practical charging data are employed to formulate an EV aggregation model, and its flexibility in both emission mitigation and energy/power dispatching is demonstrated. Second, a bilevel optimization model is formulated for EV aggregators to participate in day-ahead optimal scheduling, which aims to minimize the total cost without exceeding the given carbon cap. Third, to tackle the nonlinear coupling between the carbon flow and power flow, a bilevel model with a carbon cap constraint is formed as a constrained Markov decision process (CMDP). Finally, the CMDP is efficiently solved by the proposed augmented Lagrangian-based DRL algorithm featuring the soft actor-critic (SAC) method. Comprehensive numerical studies with IEEE distribution test feeders demonstrate that the proposed approach can achieve a fine tradeoff between cost and emission mitigation with a higher computation efficiency compared with the existing DRL methods.
科研通智能强力驱动
Strongly Powered by AbleSci AI