强化学习
汉密尔顿-雅各比-贝尔曼方程
计算机科学
时差学习
贝尔曼方程
理论(学习稳定性)
功能(生物学)
动态规划
李雅普诺夫函数
过程(计算)
人工智能
倒立摆
非线性系统
数学优化
最优控制
人工神经网络
机器学习
数学
算法
物理
操作系统
生物
进化生物学
量子力学
作者
Shixuan Yao,Xiaochen Liu,Yinghui Zhang,Ze Cui
摘要
<abstract> <p>In recent years, dynamic programming and reinforcement learning theory have been widely used to solve the nonlinear control system (NCS). Among them, many achievements have been made in the construction of network model and system stability analysis, but there is little research on establishing control strategy based on the detailed requirements of control process. Spurred by this trend, this paper proposes a detail-reward mechanism (DRM) by constructing the reward function composed of the individual detail evaluation functions in order to replace the utility function in the Hamilton-Jacobi-Bellman (HJB) equation. And this method is introduced into a wider range of deep reinforcement learning algorithms to solve optimization problems in NCS. After the mathematical description of the relevant characteristics of NCS, the stability of iterative control law is proved by Lyapunov function. With the inverted pendulum system as the experiment object, the dynamic environment is designed and the reward function is established by using the DRM. Finally, three deep reinforcement learning algorithm models are designed in the dynamic environment, which are based on Deep Q-Networks, policy gradient and actor-critic. The effects of different reward functions on the experimental accuracy are compared. The experimental results show that in NCS, using the DRM to replace the utility function in the HJB equation is more in line with the detailed requirements of the designer for the whole control process. By observing the characteristics of the system, designing the reward function and selecting the appropriate deep reinforcement learning algorithm model, the optimization problem of NCS can be solved.</p> </abstract>
科研通智能强力驱动
Strongly Powered by AbleSci AI