TY - GEN
T1 - Reinforcement Learning-Based Anti-disturbances Adaptive Control for Systems Subjected to Mismatched Disturbances and Input Uncertainties
AU - Chang, Yuxuan
AU - Zhu, Zhanxia
AU - Xing, Xiaolu
N1 - Publisher Copyright:
© 2023, Beijing HIWING Sci. and Tech. Info Inst.
PY - 2023
Y1 - 2023
N2 - This paper studies the anti-disturbances adaptive control problem with reinforcement learning (RL) actor-critic method for systems which subjected to matched, mismatched disturbances and input uncertainties. As most of the classical adaptive methods are not applicable in this case, firstly, actor-critic networks are introduced to approximate the unknown dynamics and cost function respectively. And the critic network is used to judge the performance of the actor network and give reinforcement signal to guide the updating of network weights. Furthermore, by using the hyperbolic tangent function to estimate the disturbances boundaries, the input uncertainties and time-varying disturbances can be matched and solved. As a result, an adaptive controller and a series of adaptive parameter update laws based on the backstepping method are proposed, which can accelerate the convergence under multi-source uncertainties without priori information. It also overcomes the shortcoming of data-based reinforcement learning not guaranteeing stability. Finally, through analyzing the Lyapunov function, the controller is proved to be actual exponential stable and all kinds of errors are bounded. The numerical simulation shows the validity and superiority of the proposed method.
AB - This paper studies the anti-disturbances adaptive control problem with reinforcement learning (RL) actor-critic method for systems which subjected to matched, mismatched disturbances and input uncertainties. As most of the classical adaptive methods are not applicable in this case, firstly, actor-critic networks are introduced to approximate the unknown dynamics and cost function respectively. And the critic network is used to judge the performance of the actor network and give reinforcement signal to guide the updating of network weights. Furthermore, by using the hyperbolic tangent function to estimate the disturbances boundaries, the input uncertainties and time-varying disturbances can be matched and solved. As a result, an adaptive controller and a series of adaptive parameter update laws based on the backstepping method are proposed, which can accelerate the convergence under multi-source uncertainties without priori information. It also overcomes the shortcoming of data-based reinforcement learning not guaranteeing stability. Finally, through analyzing the Lyapunov function, the controller is proved to be actual exponential stable and all kinds of errors are bounded. The numerical simulation shows the validity and superiority of the proposed method.
KW - Adaptive control
KW - Anti-disturbance control
KW - Mismatched disturbances
KW - Reinforcement learning
UR - http://www.scopus.com/inward/record.url?scp=85151051957&partnerID=8YFLogxK
U2 - 10.1007/978-981-99-0479-2_82
DO - 10.1007/978-981-99-0479-2_82
M3 - 会议稿件
AN - SCOPUS:85151051957
SN - 9789819904785
T3 - Lecture Notes in Electrical Engineering
SP - 901
EP - 910
BT - Proceedings of 2022 International Conference on Autonomous Unmanned Systems, ICAUS 2022
A2 - Fu, Wenxing
A2 - Gu, Mancang
A2 - Niu, Yifeng
PB - Springer Science and Business Media Deutschland GmbH
T2 - International Conference on Autonomous Unmanned Systems, ICAUS 2022
Y2 - 23 September 2022 through 25 September 2022
ER -