Progressive Prioritized Experience Replay for Multi-Agent Reinforcement Learning

Zhuoying Chen, Huiping Li, Rizhong Wang, Di Cui

科研成果: 书/报告/会议事项章节会议稿件同行评审

1 引用 (Scopus)

摘要

Due to the limitations of load, perception ability and communication range, single agent is difficult to meet the increasingly complex task requirements. As a result, the multi-agent reinforcement learning algorithm may attract more attention. However, the algorithm convergence becomes more difficult with the increase of the agent numbers. In this article, an efficient training framework called Progressive Prioritized Experience Replay (PPER) is proposed to resolve this problem. PPER decomposes the task scene into several similar sub-scenes with a complex degree from easy to difficult. The progressive training (PT) approach is adopted to let the agent accumulate learning experience in sub-scenes before access to the task scene, which greatly reduces the training difficulty. To verify the effectiveness of our training framework, we extended OpenAI gym to create a multi-USV confrontation environment, and the superior performance of PPER has been demonstrated in comparative tests.

源语言英语
主期刊名Proceedings of the 43rd Chinese Control Conference, CCC 2024
编辑Jing Na, Jian Sun
出版商IEEE Computer Society
8292-8296
页数5
ISBN(电子版)9789887581581
DOI
出版状态已出版 - 2024
活动43rd Chinese Control Conference, CCC 2024 - Kunming, 中国
期限: 28 7月 202431 7月 2024

出版系列

姓名Chinese Control Conference, CCC
ISSN(印刷版)1934-1768
ISSN(电子版)2161-2927

会议

会议43rd Chinese Control Conference, CCC 2024
国家/地区中国
Kunming
时期28/07/2431/07/24

指纹

探究 'Progressive Prioritized Experience Replay for Multi-Agent Reinforcement Learning' 的科研主题。它们共同构成独一无二的指纹。

引用此