Using Fuzzy Logic to Learn Abstract Policies in Large-Scale Multiagent Reinforcement Learning

Jingchen Li, Haobin Shi, Kao Shing Hwang

科研成果: 期刊稿件文章同行评审

28 引用 (Scopus)

摘要

Large-scale multiagent reinforcement learning requires huge computation and space costs, and the too-long execution process makes it hard to train policies for agents. This work proposes a concept of fuzzy agent, which is a new paradigm for training homogeneous agents. Aiming at a lightweight and affordable reinforcement learning mechanism for large-scale homogeneous multiagent systems, we break the one-to-one correspondence between agent and policy, designing abstract agents as the substitute for the multiagent to interact with the environment. The Markov decision process models for these abstract agents are conducted by fuzzy logic, which also acts on the behavior mapping from abstract agent to entity. Specifically, just the abstract agents execute their policy at a time step, and the concrete behaviors are generated by simple matrix operations. The proposal has lower space and computation complexities because the number of abstract agents is far less than that of entities, and the coupling among agents is retained implicitly. Compared with other approximation and simplification methods, the proposed fuzzy agent not only greatly reduces required computing resources but also ensures the effectiveness of the learned policies. Several experiments are conducted to validate our method. The results show that the proposal outperforms the baseline methods, while it has satisfactory zero-shot and few-shot transfer abilities.

源语言英语
页(从-至)5211-5224
页数14
期刊IEEE Transactions on Fuzzy Systems
30
12
DOI
出版状态已出版 - 1 12月 2022

指纹

探究 'Using Fuzzy Logic to Learn Abstract Policies in Large-Scale Multiagent Reinforcement Learning' 的科研主题。它们共同构成独一无二的指纹。

引用此