TY - JOUR
T1 - Actor-critic-disturbance reinforcement learning algorithm-based fast finite-time stability of multiagent systems
AU - Zhao, Junsheng
AU - Gu, Yaqi
AU - Xie, Xiangpeng
AU - Yu, Dengxiu
N1 - Publisher Copyright:
© 2024 Elsevier Inc.
PY - 2025/5
Y1 - 2025/5
N2 - This paper proposes an actor-critic-disturbance (ACD) reinforcement learning algorithm-based fast finite-time stability of multiagent systems (MASs) with time-varying asymmetrical constraints. Initially, a barrier function is designed to facilitate the transformation of the constrained system to an unconstrained one. Notably, the adaptive control strategy discussed in this paper is capable of solving more general dynamic constraints compared with most existing literature. Subsequently, in scenarios where the disturbance affects the system in the worst way, an H∞ optimal control strategy based on the ACD reinforcement learning algorithms is proposed to enhance the robustness of the system and minimize the influence of disturbances. Thirdly, a fast finite-time theory is integrated into the optimal control protocol for MASs, which allows the system to complete the control objective in finite time while converging faster. Lastly, numerical and practical simulation examples confirm the validity of the theoretical results.
AB - This paper proposes an actor-critic-disturbance (ACD) reinforcement learning algorithm-based fast finite-time stability of multiagent systems (MASs) with time-varying asymmetrical constraints. Initially, a barrier function is designed to facilitate the transformation of the constrained system to an unconstrained one. Notably, the adaptive control strategy discussed in this paper is capable of solving more general dynamic constraints compared with most existing literature. Subsequently, in scenarios where the disturbance affects the system in the worst way, an H∞ optimal control strategy based on the ACD reinforcement learning algorithms is proposed to enhance the robustness of the system and minimize the influence of disturbances. Thirdly, a fast finite-time theory is integrated into the optimal control protocol for MASs, which allows the system to complete the control objective in finite time while converging faster. Lastly, numerical and practical simulation examples confirm the validity of the theoretical results.
KW - Actor-critic-disturbance reinforcement learning
KW - Fast finite-time stabilization
KW - Time-varying asymmetrical constraint
UR - http://www.scopus.com/inward/record.url?scp=85213250617&partnerID=8YFLogxK
U2 - 10.1016/j.ins.2024.121802
DO - 10.1016/j.ins.2024.121802
M3 - 文章
AN - SCOPUS:85213250617
SN - 0020-0255
VL - 699
JO - Information Sciences
JF - Information Sciences
M1 - 121802
ER -