TY - JOUR
T1 - Mix-attention approximation for homogeneous large-scale multi-agent reinforcement learning
AU - Shike, Yang
AU - Jingchen, Li
AU - Haobin, Shi
N1 - Publisher Copyright:
© 2022, The Author(s), under exclusive licence to Springer-Verlag London Ltd., part of Springer Nature.
PY - 2023/2
Y1 - 2023/2
N2 - In large-scale multi-agent environments with homogeneous agents, most works provided approximation methods to simplify the interaction among agents. In this work, we propose a new approximation, termed mix-attention approximation, to enhance multi-agent reinforcement learning. The approximation is made by a mix-attention module, used to form consistent consensuses for agents in partially observable environments. We leverage the hard attention to compress the perception of each agent to some more partial regions. These partial regions can engage the attention of several agents at the same time, and the correlation among these partial regions is generated by a soft-attention module. We give the training method for the mix-attention mechanism and discuss the consistency between the mix-attention module and the policy network. Then we analyze the feasibility of this mix-attention-based approximation, attempting to build integrated models of our method into other approximation methods. In large-scale multi-agent environments, the proposal can be embedded into most reinforcement learning methods, and extensive experiments on multi-agent scenarios demonstrate the effectiveness of the proposed approach.
AB - In large-scale multi-agent environments with homogeneous agents, most works provided approximation methods to simplify the interaction among agents. In this work, we propose a new approximation, termed mix-attention approximation, to enhance multi-agent reinforcement learning. The approximation is made by a mix-attention module, used to form consistent consensuses for agents in partially observable environments. We leverage the hard attention to compress the perception of each agent to some more partial regions. These partial regions can engage the attention of several agents at the same time, and the correlation among these partial regions is generated by a soft-attention module. We give the training method for the mix-attention mechanism and discuss the consistency between the mix-attention module and the policy network. Then we analyze the feasibility of this mix-attention-based approximation, attempting to build integrated models of our method into other approximation methods. In large-scale multi-agent environments, the proposal can be embedded into most reinforcement learning methods, and extensive experiments on multi-agent scenarios demonstrate the effectiveness of the proposed approach.
KW - Attention mechanism
KW - Homogeneous multi-agent system
KW - Large-scale multi-agent system
KW - Reinforcement learning
UR - http://www.scopus.com/inward/record.url?scp=85139502141&partnerID=8YFLogxK
U2 - 10.1007/s00521-022-07880-4
DO - 10.1007/s00521-022-07880-4
M3 - 文章
AN - SCOPUS:85139502141
SN - 0941-0643
VL - 35
SP - 3143
EP - 3154
JO - Neural Computing and Applications
JF - Neural Computing and Applications
IS - 4
ER -