TY - JOUR
T1 - Reinforcement learning-based tracking control for a quadrotor unmanned aerial vehicle under external disturbances
AU - Liu, Hui
AU - Li, Bo
AU - Xiao, Bing
AU - Ran, Dechao
AU - Zhang, Chengxi
N1 - Publisher Copyright:
© 2022 John Wiley & Sons Ltd.
PY - 2023/11/25
Y1 - 2023/11/25
N2 - This article addresses the high-accuracy intelligent trajectory tracking control problem of a quadrotor unmanned aerial vehicle (UAV) subject to external disturbances. The tracking error systems are first reestablished by utilizing the feedforward control technique to compensate for the raw error dynamics of the quadrotor UAV. Then, two novel appointed-fixed-time observers are designed for the processed error systems to reconstruct the disturbance forces and torques, respectively. And the observation errors can converge to origin within the appointed time defined by users or designers. Subsequently, two novel control policies are developed utilizing reinforcement learning methodology, which can balance the control cost and control performance. Meanwhile, two critic neural networks are used to replace the traditional actor-critic networks for approximating the solutions of Hamilton–Jacobi–Bellman equations. More specifically, two novel weight update laws are developed. They can not only update the weights of the critic neural networks online, but also avoid utilizing the persistent excitation condition innovatively. And that the ultimately uniformly bounded stability of the whole control system is proved according to Lyapunov method by utilizing the proposed reinforcement learning-based control polices. Finally, simulation results are presented to illustrate the effectiveness and superior performances of the developed control scheme.
AB - This article addresses the high-accuracy intelligent trajectory tracking control problem of a quadrotor unmanned aerial vehicle (UAV) subject to external disturbances. The tracking error systems are first reestablished by utilizing the feedforward control technique to compensate for the raw error dynamics of the quadrotor UAV. Then, two novel appointed-fixed-time observers are designed for the processed error systems to reconstruct the disturbance forces and torques, respectively. And the observation errors can converge to origin within the appointed time defined by users or designers. Subsequently, two novel control policies are developed utilizing reinforcement learning methodology, which can balance the control cost and control performance. Meanwhile, two critic neural networks are used to replace the traditional actor-critic networks for approximating the solutions of Hamilton–Jacobi–Bellman equations. More specifically, two novel weight update laws are developed. They can not only update the weights of the critic neural networks online, but also avoid utilizing the persistent excitation condition innovatively. And that the ultimately uniformly bounded stability of the whole control system is proved according to Lyapunov method by utilizing the proposed reinforcement learning-based control polices. Finally, simulation results are presented to illustrate the effectiveness and superior performances of the developed control scheme.
KW - adaptive dynamic programming
KW - appointed-fixed-time observer
KW - reinforcement learning
KW - trajectory tracking control
KW - unmanned aerial vehicle
UR - http://www.scopus.com/inward/record.url?scp=85135510771&partnerID=8YFLogxK
U2 - 10.1002/rnc.6334
DO - 10.1002/rnc.6334
M3 - 文章
AN - SCOPUS:85135510771
SN - 1049-8923
VL - 33
SP - 10360
EP - 10377
JO - International Journal of Robust and Nonlinear Control
JF - International Journal of Robust and Nonlinear Control
IS - 17
ER -