Transfer reinforcement learning for multi-agent pursuit-evasion differential game with obstacles in a continuous environment

Penglin Hu, Quan Pan, Chunhui Zhao, Yaning Guo

Research output: Contribution to journalArticlepeer-review

3 Scopus citations

Abstract

In this paper, we study the multi-pursuer single-evader pursuit-evasion (MSPE) differential game in a continuous environment with the consideration of obstacles. We propose a novel pursuit-evasion algorithm based on reinforcement learning and transfer learning. In the source task learning stage, we employ the Q-learning and value function approximation method to overcome the challenges posed by the large-scale storage space required by the conventional Q-table solution method. This approach expands the discrete space to the continuous space by value function approximation and effectively reduces the demand for storage space. During the target task learning stage, we utilize the Gaussian mixture model (GMM) to classify the source tasks. The source policies whose corresponding state-value sets have the highest probability densities are assigned for the agent in the target task for learning. This methodology not only effectively avoids negative transfer but also enhances the algorithm's generalization ability and convergence speed. Through simulation and experiment, we demonstrate the algorithm's effectiveness.

Original languageEnglish
Pages (from-to)2125-2140
Number of pages16
JournalAsian Journal of Control
Volume26
Issue number4
DOIs
StatePublished - Jul 2024

Keywords

  • Gaussian mixture model
  • Q-learning
  • pursuit-evasion game
  • recursive least squares
  • transfer reinforcement learning
  • value function approximation

Fingerprint

Dive into the research topics of 'Transfer reinforcement learning for multi-agent pursuit-evasion differential game with obstacles in a continuous environment'. Together they form a unique fingerprint.

Cite this