Multi-vehicle Flocking Control with Deep Deterministic Policy Gradient Method

科研成果: 书/报告/会议事项章节会议稿件同行评审

35 引用 (Scopus)

摘要

Flocking control has been studied extensively along with the wide application of multi-vehicle systems. In this paper the Multi-vehicles System (MVS) flocking control with collision avoidance and communication preserving is considered based on the deep reinforcement learning framework. Specifically the deep deterministic policy gradient (DDPG) with centralized training and distributed execution process is implemented to obtain the flocking control policy. First, to avoid the dynamically changed observation of state, a three layers tensor based representation of the observation is used so that the state remains constant although the observation dimension is changing. A reward function is designed to guide the way-points tracking, collision avoidance and communication preserving. The reward function is augmented by introducing the local reward function of neighbors. Finally, a centralized training process which trains the shared policy based on common training set among all agents. The proposed method is tested under simulated scenarios with different setup.

源语言英语
主期刊名2018 IEEE 14th International Conference on Control and Automation, ICCA 2018
出版商IEEE Computer Society
306-311
页数6
ISBN(印刷版)9781538660898
DOI
出版状态已出版 - 21 8月 2018
活动14th IEEE International Conference on Control and Automation, ICCA 2018 - Anchorage, 美国
期限: 12 6月 201815 6月 2018

出版系列

姓名IEEE International Conference on Control and Automation, ICCA
2018-June
ISSN(印刷版)1948-3449
ISSN(电子版)1948-3457

会议

会议14th IEEE International Conference on Control and Automation, ICCA 2018
国家/地区美国
Anchorage
时期12/06/1815/06/18

指纹

探究 'Multi-vehicle Flocking Control with Deep Deterministic Policy Gradient Method' 的科研主题。它们共同构成独一无二的指纹。

引用此