TY - GEN
T1 - Enhancing Simultaneous Arrival at a Dynamic Target
T2 - 18th International Conference on Control, Automation, Robotics and Vision, ICARCV 2024
AU - Lei, Yifei
AU - Hu, Jinwen
AU - Xu, Zhao
AU - Gao, Chenqi
AU - Li, Jiatong
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - This paper addresses the complex challenge of ensuring simultaneous arrival of multiple agents at a dynamic target, a critical requirement for operations such as roundups and saturation attacks. Traditional guidance systems have faced significant hurdles due to inaccurate flight time estimates and difficulties adapting to high-speed maneuvers. To overcome these limitations, we introduce a novel framework that utilizes Proximal Policy Optimization (PPO) and expert rules. This approach leverages distributed computing to enable autonomous decision-making among agents, thereby simplifying the deep reinforcement learning model and reducing computational overhead, which enhances scalability and adaptability. Additionally, we incorporate an angular velocity reward into the reward function, improving the predictability and effectiveness of maneuvers, particularly for targets with high-speed and irregular trajectories. The proposed methods have been rigorously tested through numerous simulations and high-fidelity scenarios, confirming their robustness and superior performance over traditional and enhanced proportional guidance systems.
AB - This paper addresses the complex challenge of ensuring simultaneous arrival of multiple agents at a dynamic target, a critical requirement for operations such as roundups and saturation attacks. Traditional guidance systems have faced significant hurdles due to inaccurate flight time estimates and difficulties adapting to high-speed maneuvers. To overcome these limitations, we introduce a novel framework that utilizes Proximal Policy Optimization (PPO) and expert rules. This approach leverages distributed computing to enable autonomous decision-making among agents, thereby simplifying the deep reinforcement learning model and reducing computational overhead, which enhances scalability and adaptability. Additionally, we incorporate an angular velocity reward into the reward function, improving the predictability and effectiveness of maneuvers, particularly for targets with high-speed and irregular trajectories. The proposed methods have been rigorously tested through numerous simulations and high-fidelity scenarios, confirming their robustness and superior performance over traditional and enhanced proportional guidance systems.
UR - http://www.scopus.com/inward/record.url?scp=85217399205&partnerID=8YFLogxK
U2 - 10.1109/ICARCV63323.2024.10821659
DO - 10.1109/ICARCV63323.2024.10821659
M3 - 会议稿件
AN - SCOPUS:85217399205
T3 - 2024 18th International Conference on Control, Automation, Robotics and Vision, ICARCV 2024
SP - 43
EP - 48
BT - 2024 18th International Conference on Control, Automation, Robotics and Vision, ICARCV 2024
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 12 December 2024 through 15 December 2024
ER -