PMIC: Improving Multi-Agent Reinforcement Learning with Progressive Mutual Information Collaboration

Pengyi Li, Hongyao Tang, Tianpei Yang, Xiaotian Hao, Tong Sang, Yan Zheng, Jianye Hao, Matthew E. Taylor, Wenyuan Tao, Zhen Wang

科研成果: 期刊稿件会议文章同行评审

20 引用 (Scopus)

摘要

Learning to collaborate is critical in MultiAgent Reinforcement Learning (MARL). Previous works promote collaboration by maximizing the correlation of agents' behaviors, which is typically characterized by Mutual Information (MI) in different forms. However, we reveal sub-optimal collaborative behaviors also emerge with strong correlations, and simply maximizing the MI can, surprisingly, hinder the learning towards better collaboration. To address this issue, we propose a novel MARL framework, called Progressive Mutual Information Collaboration (PMIC), for more effective MI-driven collaboration. PMIC uses a new collaboration criterion measured by the MI between global states and joint actions. Based on this criterion, the key idea of PMIC is maximizing the MI associated with superior collaborative behaviors and minimizing the MI associated with inferior ones. The two MI objectives play complementary roles by facilitating better collaborations while avoiding falling into sub-optimal ones. Experiments on a wide range of MARL benchmarks show the superior performance of PMIC compared with other algorithms.

源语言英语
页(从-至)12979-12997
页数19
期刊Proceedings of Machine Learning Research
162
出版状态已出版 - 2022
活动39th International Conference on Machine Learning, ICML 2022 - Baltimore, 美国
期限: 17 7月 202223 7月 2022

指纹

探究 'PMIC: Improving Multi-Agent Reinforcement Learning with Progressive Mutual Information Collaboration' 的科研主题。它们共同构成独一无二的指纹。

引用此