Treffer: MALib: A Parallel Framework for Population-based Multi-agent Reinforcement Learning.

Title:
MALib: A Parallel Framework for Population-based Multi-agent Reinforcement Learning.
Authors:
Ming Zhou1 MINGAK@SJTU.EDU.CN, Ziyu Wan1 ALEX_WAN@SJTU.EDU.CN, Hanjing Wang1 WANGHANJINGWHJ@SJTU.EDU.CN, Muning Wen1 MUNING.WEN@OUTLOOK.COM, Runzhe Wu1 RUNZHE@SJTU.EDU.CN, Ying Wen1 YING.WEN@SJTU.EDU.CN, Yaodong Yang2 YAODONG.YANG@PKU.EDU.CN, Yong Yu1 YYU@APEX.SJTU.EDU.CN, Jun Wang3 JUN.WANG@CS.UCL.AC.UK, Weinan Zhang1 WNZHANG@SJTU.COM
Source:
Journal of Machine Learning Research. 2023, Vol. 24, p1-12. 12p.
Database:
Academic Search Index

Weitere Informationen

Population-based multi-agent reinforcement learning (PB-MARL) encompasses a range of methods that merge dynamic population selection with multi-agent reinforcement learning algorithms (MARL). While PB-MARL has demonstrated notable achievements in complex multi-agent tasks, its sequential execution is plagued by low computational efficiency due to the diversity in computing patterns and policy combinations. We propose a solution involving a stateless central task dispatcher and stateful workers to handle PB-MARL's subroutines, thereby capitalizing on parallelism across various components for efficient problem-solving. In line with this approach, we introduce MALib, a parallel framework that incorporates a task control model, independent data servers, and an abstraction of MARL training paradigms. The framework has undergone extensive testing and is available under the MIT license (https://github.com/sjtu-marl/malib). [ABSTRACT FROM AUTHOR]