线性二次调节器
有界函数
稳健性(进化)
数学
控制理论(社会学)
最优控制
线性系统
数学优化
理论(学习稳定性)
应用数学
计算机科学
控制(管理)
数学分析
基因
生物化学
机器学习
人工智能
化学
作者
Bo Pang,Tao Bian,Zhong‐Ping Jiang
标识
DOI:10.1109/tac.2021.3085510
摘要
This article studies the robustness of policy iteration in the context of continuous-time infinite-horizon linear quadratic regulator (LQR) problem. It is shown that Kleinman's policy iteration algorithm is small-disturbance input-to-state stable, a property that is stronger than Sontag's local input-to-state stability but weaker than global input-to-state stability. More precisely, whenever the error in each iteration is bounded and small, the solutions of the policy iteration algorithm are also bounded and enter a small neighborhood of the optimal solution of the LQR problem. Based on this result, an off-policy data-driven policy iteration algorithm for the LQR problem is shown to be robust when the system dynamics are subject to small additive unknown bounded disturbances. The theoretical results are validated by a numerical example.
科研通智能强力驱动
Strongly Powered by AbleSci AI