Title
Qauxi: Cooperative multi-agent reinforcement learning with knowledge transferred from auxiliary task
Abstract
Deep multi-agent reinforcement learning (MARL) can efficiently learn decentralized policies for real-world applications. However, current MARL methods suffer from the difficulty of transferring knowledge from already learned tasks to improve its exploration. In this paper, we propose a novel MARL method called Qauxi, which forms coordinated exploration scheme to improve the traditional MARL algorithms by reusing the meta-experience transferred from auxiliary task. We also use the weighting function to weight the importance of the joint action in monotonic loss function in order to focus on more important joint actions and thus avoid yielding suboptimal policies. Furthermore, we prove the convergence of Qauxi based on contraction mapping theorem. Qauxi is evaluated on the widely adopted StarCraft benchmarks (SMAC) across easy, hard, and super hard scenarios. Experimental results show that the proposed method outperforms the state-of-the-art MARL methods by a large margin in the most challenging super hard scenarios.
Year
DOI
Venue
2022
10.1016/j.neucom.2022.06.091
Neurocomputing
Keywords
DocType
Volume
Multi-agent,Transfer learning,Reinforcement learning,MARL,Starcraft
Journal
504
ISSN
Citations 
PageRank 
0925-2312
0
0.34
References 
Authors
0
7
Name
Order
Citations
PageRank
Wenqian Liang100.34
Ji Wang219036.75
Weidong Bao3236.49
Xiaomin Zhu4921100.31
Guanlin Wu500.34
Dayu Zhang600.34
Liyuan Niu700.34