ME-MADDPG: An efficient learning-based motion planning method for multiple agents in complex environments

Kaifang Wan, Dingwei Wu, Bo Li, Xiaoguang Gao, Zijian Hu, Daqing Chen

科研成果: 期刊稿件文章同行评审

38 引用 (Scopus)

摘要

Developing efficient motion policies for multiagents is a challenge in a decentralized dynamic situation, where each agent plans its own paths without knowing the policies of the other agents involved. This paper presents an efficient learning-based motion planning method for multiagent systems. It adopts the framework of multiagent deep deterministic policy gradient (MADDPG) to directly map partially observed information to motion commands for multiple agents. To improve the efficiency of MADDPG in sample utilization, so as to train more brilliant agents that can adapt to more complex environments, a strategy named mixed experience (ME) is introduced to MADDPG, and this has led to our proposed ME-MADDPG algorithm. The novel ME strategy can be embodied into three specific mechanisms: (1) an artificial potential field-based sample generator to produce high-quality samples in the early training stage; (2) a dynamic mixed sampling strategy to mix the training data from different sources with a variable proportion; (3) a delayed learning skill to stabilize the training of the multiple agents. A series of experiments have been conducted to verify the performance of the proposed ME-MADDPG algorithm, and it has been demonstrated that, compared with MADDPG, the proposed algorithm can significantly improve the convergence speed and convergence effect in the training process, and it has also shown better efficiency and better adaptability in complex dynamic environments while it is used for multiagent motion planning applications.

源语言英语
页(从-至)2393-2427
页数35
期刊International Journal of Intelligent Systems
37
3
DOI
出版状态已出版 - 3月 2022

指纹

探究 'ME-MADDPG: An efficient learning-based motion planning method for multiple agents in complex environments' 的科研主题。它们共同构成独一无二的指纹。

引用此