1. K. Kurach, A. Raichuk, P. Stańczyk, M. Zając, O. Bachem, L. Espeholt, C. Riquehne, D. Vincent, M. Michalski, O. Bousquet, S. Gelly. Google research football: A novel reinforcement learning environment. In Proceedings of the 34th AAAI Conference on Artificial Intelligence, New York, USA, pp.4501-4510, 2020. DOI: https://doi.org/10.1609/aaai.v34i04.5878.
2. C. S. de Witt, T. Gupta, D. Makoviichuk, V. Makoviychuk, P. H. S. Torr, M. F. Sun, S. Whiteson. Is independent learning all you need in the StarCraft multi-agent challenge? [Online], Available: https://arxiv.org/abs/2011.09533, 2020.
3. M. Zhou, Z. Y. Wan, H. J. Wang, M. N. Wen, R. Z. Wu, Y. Wen, Y. D. Yang, W. N. Zhang, J. Wang. MALib: A paraUel framework for population-based multi-agent reinforcement learning. [Online], Available: https://arxiv.org/abs/2106.07551, 2021.
4. M. Jaderberg, W. M. Czarnecki, I. Dunning, L. Marris, G. Lever, A. G. Castaneda, C. Beattie, N. C. Rabinowitz, A. S. Morcos, A. Ruderman, N. Sonnerat, T. Green, L. Deason, J. Z. Leibo, D. Silver, D. Hassabis, K. Kavukcuoglu, T. Graepel. Human-level performance in first-person multiplayer games with population-based deep reinforcement learning. [Online], Available: https://arxiv.org/abs/1807.01281, 2018.
5. J. Schrittwieser, I. Antonoglou, T. Hubert, K. Simonyan, L. Sifre, S. Schmitt, A. Guez, E. Lockhart, D. Hassabis, T. Graepel, T. Lillicrap, D. Silver. Mastering Atari, Go, chess and Shogi by planning with a learned model. Nature, vol. 588, no. 7839, pp. 604–609, 2020. DOI: https://doi.org/10.1038/S41586-020-03051-4.