A Fuzzy Curiosity-Driven Mechanism for Multi-Agent Reinforcement Learning

Wenbai Chen, Haobin Shi, Jingchen Li, Kao Shing Hwang

Research output: Contribution to journalArticlepeer-review

8 Scopus citations

Abstract

Many works provide intrinsic rewards to deal with sparse rewards in reinforcement learning. Due to the non-stationarity of multi-agent systems, it is impracticable to apply existing methods to multi-agent reinforcement learning directly. In this paper, a fuzzy curiosity-driven mechanism is proposed for multi-agent reinforcement learning, by which agents can explore more efficiently in a scenario with sparse extrinsic reward. First, we improve the variational auto-encoder to predict the next state through the joint-state and joint-action for agents. Then several fuzzy partitions are built according to the next joint-state, which aims at assigning the prediction error to different agents. With the proposed method, each agent in the multi-agent environment receives its individual intrinsic reward. We elaborate on the proposed method in partially observable environments and fully observable environments separately. Experimental results show that multi-agent learns joint policies more efficiently by the proposed fuzzy curiosity-driven mechanism, and it can also help agents find better policies in the training process.

Original languageEnglish
Pages (from-to)1222-1233
Number of pages12
JournalInternational Journal of Fuzzy Systems
Volume23
Issue number5
DOIs
StatePublished - Jul 2021
Externally publishedYes

Keywords

  • CURIOSITY-driven
  • Multi-agent system
  • Reinforcement learning

Fingerprint

Dive into the research topics of 'A Fuzzy Curiosity-Driven Mechanism for Multi-Agent Reinforcement Learning'. Together they form a unique fingerprint.

Cite this