1. Amato, C., Konidaris, G., Cruz, G., Maynor, C., How, J., Kaelbling, L.: Planning for decentralized control of multiple robots under uncertainty 2015, 5 (2015)
2. Aubret, A., Matignon, L., Hassas, S.: A survey on intrinsic motivation in reinforcement learning (2019)
3. Bellemare, M.G., Naddaf, Y., Veness, J., Bowling, M.: The arcade learning environment: an evaluation platform for general agents. J. Artif. Intell. Res. 47, 253–279 (2013)
4. Bellemare, M., Srinivasan, S., Ostrovski, G., Schaul, T., Saxton, D., Munos, R.: Unifying count-based exploration and intrinsic motivation. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc. (2016)
5. Bellemare, M., Veness, J., Talvitie, E.: Skip context tree switching. In: Xing, E.P., Jebara, T. (eds.) Proceedings of the 31st International Conference on Machine Learning, volume 32 of Proceedings of Machine Learning Research, pp. 1458–1466, Bejing, China, 22–24 Jun 2014. PMLR