The survey shows:
Despite the fast development of multi-agent reinforcement learning (MARL) algorithms, there is a lack of commonly-acknowledged baseline implementation and evaluation benchmarks.
An urgent need for MARL researchers is to develop a unified benchmark suite, similar to the role of RLlib in single-agent RL, that can support both high-performance MARL implementations and replicable evaluations in various testing environments.