A collaboration of multi-agent model using an interactive interface

Jingchen Li, Fan Wu, Haobin Shi, Kao Shing Hwang

Research output: Contribution to journalArticlepeer-review

5 Scopus citations

Abstract

Multi-agent reinforcement learning algorithms scarcely attend to noisy environments, in which agents are inhibited from achieving optimal policy training and making correct decisions. This work investigates the effect of noises in multi-agent environments and proposes a multi-agent actor-critic with collaboration (MACC) model. The model uses lightweight communication to overcome the interference from noises. There are two policies for each agent in MACC: collaboration policy and behavior policy. The behavior of an agent not only depends on its own state but also be influenced by each other agent through a scalar, collaboration value. The collaboration value is generated by the collaboration policy for each individual agent, and it ensures a succinct consensus about the environment. This paper elaborates on the training of the collaboration policy and specifies how it coordinates the behavior policy in a manner of temporal abstraction mechanism, while the observation sequence is considered for more accurate perception. Several experiments on multi-agent collaboration simulation platforms demonstrate that the MACC performs better than baselines in noisy environments, especially in partially observable environments.

Original languageEnglish
Pages (from-to)349-363
Number of pages15
JournalInformation Sciences
Volume611
DOIs
StatePublished - Sep 2022

Keywords

  • Actor-critic
  • Multi-agent reinforcement learning
  • Partial observable environment

Fingerprint

Dive into the research topics of 'A collaboration of multi-agent model using an interactive interface'. Together they form a unique fingerprint.

Cite this