控制理论(社会学)
强化学习
自适应控制
李雅普诺夫函数
控制器(灌溉)
非线性系统
Lyapunov稳定性
计算机科学
理论(学习稳定性)
迭代学习控制
弹道
控制系统
工程类
控制(管理)
人工智能
物理
电气工程
量子力学
机器学习
天文
农学
生物
作者
Anuradha M. Annaswamy,Anubhav Guha,Yingnan Cui,Sunbochen Tang,Peter Fisher,Joseph E. Gaudio
标识
DOI:10.1109/tac.2023.3290037
摘要
This article considers the problem of real-time control and learning in dynamic systems subjected to parameteric uncertainties. We propose a combination of a reinforcement learning (RL)-based policy in the outer loop suitably chosen to ensure stability and optimality for the nominal dynamics, together with adaptive control (AC) in the inner loop so that in real-time AC contracts the closed-loop dynamics toward a stable trajectory traced out by RL. In total, two classes of nonlinear dynamic systems are considered, both of which are control affine. The first class of dynamic systems utilizes equilibrium points and a Lyapunov approach, whereas second class of nonlinear systems uses contraction theory. AC-RL controllers are proposed for both classes of systems and shown to lead to online policies that guarantee stability using a high-order tuner and accommodate parameteric uncertainties and magnitude limits on the input. In addition to establishing a stability guarantee with real-time control, the AC-RL controller is also shown to lead to parameter learning with persistent excitation for the first class of systems. Numerical validations of all algorithms are carried out using a quadrotor landing task on a moving platform.
科研通智能强力驱动
Strongly Powered by AbleSci AI